Sep 29 19:09:17 crc systemd[1]: Starting Kubernetes Kubelet... Sep 29 19:09:17 crc restorecon[4736]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:17 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 19:09:18 crc restorecon[4736]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Sep 29 19:09:18 crc kubenswrapper[4741]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 19:09:18 crc kubenswrapper[4741]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Sep 29 19:09:18 crc kubenswrapper[4741]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 19:09:18 crc kubenswrapper[4741]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 19:09:18 crc kubenswrapper[4741]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 29 19:09:18 crc kubenswrapper[4741]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.797222 4741 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805298 4741 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805331 4741 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805341 4741 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805349 4741 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805360 4741 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805371 4741 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805379 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805392 4741 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805423 4741 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805431 4741 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805439 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805468 4741 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805476 4741 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805484 4741 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805492 4741 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805499 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805510 4741 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805519 4741 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805529 4741 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805538 4741 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805547 4741 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805555 4741 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805563 4741 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805571 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805579 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805587 4741 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805594 4741 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805602 4741 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805610 4741 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805617 4741 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805625 4741 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805632 4741 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805639 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805648 4741 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805656 4741 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805663 4741 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805671 4741 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805678 4741 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805686 4741 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805693 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805701 4741 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805708 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805716 4741 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805724 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805732 4741 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805740 4741 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805753 4741 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805762 4741 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805771 4741 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805779 4741 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805788 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805796 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805804 4741 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805812 4741 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805822 4741 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805831 4741 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805839 4741 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805847 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805854 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805862 4741 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805870 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805878 4741 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805886 4741 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805893 4741 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805901 4741 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805908 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805916 4741 feature_gate.go:330] unrecognized feature gate: Example Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805924 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805931 4741 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805941 4741 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.805949 4741 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807169 4741 flags.go:64] FLAG: --address="0.0.0.0" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807196 4741 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807210 4741 flags.go:64] FLAG: --anonymous-auth="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807222 4741 flags.go:64] FLAG: --application-metrics-count-limit="100" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807238 4741 flags.go:64] FLAG: --authentication-token-webhook="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807249 4741 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807264 4741 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807277 4741 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807289 4741 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807301 4741 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807314 4741 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807329 4741 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807340 4741 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807351 4741 flags.go:64] FLAG: --cgroup-root="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807361 4741 flags.go:64] FLAG: --cgroups-per-qos="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807373 4741 flags.go:64] FLAG: --client-ca-file="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807383 4741 flags.go:64] FLAG: --cloud-config="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807444 4741 flags.go:64] FLAG: --cloud-provider="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807465 4741 flags.go:64] FLAG: --cluster-dns="[]" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807615 4741 flags.go:64] FLAG: --cluster-domain="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807629 4741 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807640 4741 flags.go:64] FLAG: --config-dir="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807650 4741 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807663 4741 flags.go:64] FLAG: --container-log-max-files="5" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807679 4741 flags.go:64] FLAG: --container-log-max-size="10Mi" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807690 4741 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807705 4741 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807717 4741 flags.go:64] FLAG: --containerd-namespace="k8s.io" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807729 4741 flags.go:64] FLAG: --contention-profiling="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807741 4741 flags.go:64] FLAG: --cpu-cfs-quota="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807752 4741 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807764 4741 flags.go:64] FLAG: --cpu-manager-policy="none" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807773 4741 flags.go:64] FLAG: --cpu-manager-policy-options="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807785 4741 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807794 4741 flags.go:64] FLAG: --enable-controller-attach-detach="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807803 4741 flags.go:64] FLAG: --enable-debugging-handlers="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807812 4741 flags.go:64] FLAG: --enable-load-reader="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807821 4741 flags.go:64] FLAG: --enable-server="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807830 4741 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807840 4741 flags.go:64] FLAG: --event-burst="100" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807850 4741 flags.go:64] FLAG: --event-qps="50" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807859 4741 flags.go:64] FLAG: --event-storage-age-limit="default=0" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807868 4741 flags.go:64] FLAG: --event-storage-event-limit="default=0" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807877 4741 flags.go:64] FLAG: --eviction-hard="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807888 4741 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807897 4741 flags.go:64] FLAG: --eviction-minimum-reclaim="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807906 4741 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807915 4741 flags.go:64] FLAG: --eviction-soft="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807924 4741 flags.go:64] FLAG: --eviction-soft-grace-period="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807932 4741 flags.go:64] FLAG: --exit-on-lock-contention="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807941 4741 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807950 4741 flags.go:64] FLAG: --experimental-mounter-path="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807959 4741 flags.go:64] FLAG: --fail-cgroupv1="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807967 4741 flags.go:64] FLAG: --fail-swap-on="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807976 4741 flags.go:64] FLAG: --feature-gates="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807987 4741 flags.go:64] FLAG: --file-check-frequency="20s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.807996 4741 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808006 4741 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808015 4741 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808024 4741 flags.go:64] FLAG: --healthz-port="10248" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808033 4741 flags.go:64] FLAG: --help="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808042 4741 flags.go:64] FLAG: --hostname-override="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808052 4741 flags.go:64] FLAG: --housekeeping-interval="10s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808061 4741 flags.go:64] FLAG: --http-check-frequency="20s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808070 4741 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808078 4741 flags.go:64] FLAG: --image-credential-provider-config="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808087 4741 flags.go:64] FLAG: --image-gc-high-threshold="85" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808096 4741 flags.go:64] FLAG: --image-gc-low-threshold="80" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808105 4741 flags.go:64] FLAG: --image-service-endpoint="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808113 4741 flags.go:64] FLAG: --kernel-memcg-notification="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808122 4741 flags.go:64] FLAG: --kube-api-burst="100" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808131 4741 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808140 4741 flags.go:64] FLAG: --kube-api-qps="50" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808149 4741 flags.go:64] FLAG: --kube-reserved="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808158 4741 flags.go:64] FLAG: --kube-reserved-cgroup="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808166 4741 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808175 4741 flags.go:64] FLAG: --kubelet-cgroups="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808184 4741 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808193 4741 flags.go:64] FLAG: --lock-file="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808202 4741 flags.go:64] FLAG: --log-cadvisor-usage="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808210 4741 flags.go:64] FLAG: --log-flush-frequency="5s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808219 4741 flags.go:64] FLAG: --log-json-info-buffer-size="0" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808232 4741 flags.go:64] FLAG: --log-json-split-stream="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808242 4741 flags.go:64] FLAG: --log-text-info-buffer-size="0" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808251 4741 flags.go:64] FLAG: --log-text-split-stream="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808259 4741 flags.go:64] FLAG: --logging-format="text" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808268 4741 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808277 4741 flags.go:64] FLAG: --make-iptables-util-chains="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808286 4741 flags.go:64] FLAG: --manifest-url="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808295 4741 flags.go:64] FLAG: --manifest-url-header="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808309 4741 flags.go:64] FLAG: --max-housekeeping-interval="15s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808319 4741 flags.go:64] FLAG: --max-open-files="1000000" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808333 4741 flags.go:64] FLAG: --max-pods="110" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808354 4741 flags.go:64] FLAG: --maximum-dead-containers="-1" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808374 4741 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808386 4741 flags.go:64] FLAG: --memory-manager-policy="None" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808439 4741 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808450 4741 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808462 4741 flags.go:64] FLAG: --node-ip="192.168.126.11" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808471 4741 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808511 4741 flags.go:64] FLAG: --node-status-max-images="50" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808526 4741 flags.go:64] FLAG: --node-status-update-frequency="10s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808538 4741 flags.go:64] FLAG: --oom-score-adj="-999" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808550 4741 flags.go:64] FLAG: --pod-cidr="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808562 4741 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808583 4741 flags.go:64] FLAG: --pod-manifest-path="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808594 4741 flags.go:64] FLAG: --pod-max-pids="-1" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808606 4741 flags.go:64] FLAG: --pods-per-core="0" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808618 4741 flags.go:64] FLAG: --port="10250" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808629 4741 flags.go:64] FLAG: --protect-kernel-defaults="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808641 4741 flags.go:64] FLAG: --provider-id="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808652 4741 flags.go:64] FLAG: --qos-reserved="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808663 4741 flags.go:64] FLAG: --read-only-port="10255" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808675 4741 flags.go:64] FLAG: --register-node="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808687 4741 flags.go:64] FLAG: --register-schedulable="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808696 4741 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808715 4741 flags.go:64] FLAG: --registry-burst="10" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808726 4741 flags.go:64] FLAG: --registry-qps="5" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808738 4741 flags.go:64] FLAG: --reserved-cpus="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808749 4741 flags.go:64] FLAG: --reserved-memory="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808762 4741 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808776 4741 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808788 4741 flags.go:64] FLAG: --rotate-certificates="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808800 4741 flags.go:64] FLAG: --rotate-server-certificates="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808811 4741 flags.go:64] FLAG: --runonce="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808822 4741 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808834 4741 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808846 4741 flags.go:64] FLAG: --seccomp-default="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808857 4741 flags.go:64] FLAG: --serialize-image-pulls="true" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808869 4741 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808881 4741 flags.go:64] FLAG: --storage-driver-db="cadvisor" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808891 4741 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808902 4741 flags.go:64] FLAG: --storage-driver-password="root" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808913 4741 flags.go:64] FLAG: --storage-driver-secure="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808926 4741 flags.go:64] FLAG: --storage-driver-table="stats" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808938 4741 flags.go:64] FLAG: --storage-driver-user="root" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808949 4741 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808961 4741 flags.go:64] FLAG: --sync-frequency="1m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808973 4741 flags.go:64] FLAG: --system-cgroups="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.808984 4741 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809003 4741 flags.go:64] FLAG: --system-reserved-cgroup="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809015 4741 flags.go:64] FLAG: --tls-cert-file="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809026 4741 flags.go:64] FLAG: --tls-cipher-suites="[]" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809041 4741 flags.go:64] FLAG: --tls-min-version="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809053 4741 flags.go:64] FLAG: --tls-private-key-file="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809064 4741 flags.go:64] FLAG: --topology-manager-policy="none" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809074 4741 flags.go:64] FLAG: --topology-manager-policy-options="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809083 4741 flags.go:64] FLAG: --topology-manager-scope="container" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809092 4741 flags.go:64] FLAG: --v="2" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809106 4741 flags.go:64] FLAG: --version="false" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809135 4741 flags.go:64] FLAG: --vmodule="" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809150 4741 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.809163 4741 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809369 4741 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809384 4741 feature_gate.go:330] unrecognized feature gate: Example Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809442 4741 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809456 4741 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809467 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809478 4741 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809490 4741 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809499 4741 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809509 4741 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809521 4741 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809531 4741 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809541 4741 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809551 4741 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809561 4741 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809574 4741 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809586 4741 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809598 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809611 4741 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809622 4741 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809632 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809642 4741 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809651 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809661 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809671 4741 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809681 4741 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809689 4741 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809698 4741 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809705 4741 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809714 4741 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809722 4741 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809729 4741 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809736 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809744 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809752 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809760 4741 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809768 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809776 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809783 4741 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809791 4741 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809798 4741 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809806 4741 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809813 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809824 4741 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809832 4741 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809841 4741 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809849 4741 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809856 4741 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809864 4741 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809872 4741 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809879 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809887 4741 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809895 4741 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809903 4741 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809911 4741 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809919 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809927 4741 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809935 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809942 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809950 4741 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809957 4741 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809965 4741 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809973 4741 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809981 4741 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809989 4741 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.809997 4741 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.810005 4741 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.810013 4741 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.810020 4741 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.810028 4741 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.810036 4741 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.810043 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.811017 4741 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.821632 4741 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.821658 4741 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821727 4741 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821734 4741 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821740 4741 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821747 4741 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821751 4741 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821755 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821759 4741 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821762 4741 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821767 4741 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821772 4741 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821776 4741 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821780 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821784 4741 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821788 4741 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821791 4741 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821795 4741 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821799 4741 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821802 4741 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821806 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821809 4741 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821813 4741 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821817 4741 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821820 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821824 4741 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821829 4741 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821834 4741 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821838 4741 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821842 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821846 4741 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821851 4741 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821855 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821859 4741 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821863 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821867 4741 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821871 4741 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821875 4741 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821878 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821882 4741 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821886 4741 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821889 4741 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821893 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821896 4741 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821900 4741 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821904 4741 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821907 4741 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821911 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821914 4741 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821918 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821921 4741 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821925 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821928 4741 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821932 4741 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821938 4741 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821942 4741 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821946 4741 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821949 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821953 4741 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821956 4741 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821960 4741 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821963 4741 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821967 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821970 4741 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821975 4741 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821979 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821983 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821986 4741 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821990 4741 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821995 4741 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.821999 4741 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822003 4741 feature_gate.go:330] unrecognized feature gate: Example Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822008 4741 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.822013 4741 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822130 4741 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822138 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822142 4741 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822146 4741 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822150 4741 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822153 4741 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822157 4741 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822161 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822164 4741 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822168 4741 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822171 4741 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822175 4741 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822178 4741 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822182 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822186 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822191 4741 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822196 4741 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822200 4741 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822204 4741 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822220 4741 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822224 4741 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822228 4741 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822231 4741 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822237 4741 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822241 4741 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822246 4741 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822250 4741 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822254 4741 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822258 4741 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822262 4741 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822266 4741 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822270 4741 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822274 4741 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822278 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822281 4741 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822285 4741 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822290 4741 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822293 4741 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822297 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822300 4741 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822304 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822307 4741 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822311 4741 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822315 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822318 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822322 4741 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822326 4741 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822329 4741 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822332 4741 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822336 4741 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822339 4741 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822343 4741 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822346 4741 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822350 4741 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822353 4741 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822357 4741 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822377 4741 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822381 4741 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822386 4741 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822393 4741 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822408 4741 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822412 4741 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822417 4741 feature_gate.go:330] unrecognized feature gate: Example Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822421 4741 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822425 4741 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822430 4741 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822434 4741 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822437 4741 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822441 4741 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822446 4741 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 19:09:18 crc kubenswrapper[4741]: W0929 19:09:18.822451 4741 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.822456 4741 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.823980 4741 server.go:940] "Client rotation is on, will bootstrap in background" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.827318 4741 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.827416 4741 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.829568 4741 server.go:997] "Starting client certificate rotation" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.829595 4741 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.829851 4741 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-17 03:48:16.695682581 +0000 UTC Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.829987 4741 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1160h38m57.865702193s for next certificate rotation Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.872780 4741 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.876463 4741 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.902130 4741 log.go:25] "Validated CRI v1 runtime API" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.970034 4741 log.go:25] "Validated CRI v1 image API" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.971671 4741 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.980938 4741 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-09-29-19-04-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Sep 29 19:09:18 crc kubenswrapper[4741]: I0929 19:09:18.980969 4741 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:18.999943 4741 manager.go:217] Machine: {Timestamp:2025-09-29 19:09:18.995383073 +0000 UTC m=+0.643172415 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:43f60388-9284-4af8-8654-468a0103453b BootID:8fb5d538-81ea-402d-ba81-35b3f9382fe2 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7e:bf:de Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:7e:bf:de Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:0b:d3:52 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ba:02:f6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:45:50:b9 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:cc:4f:4f Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:e6:f7:b2 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:e0:58:18:15:0d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:12:8e:00:c9:72:8f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.000180 4741 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.000379 4741 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.000886 4741 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.001059 4741 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.001094 4741 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.001308 4741 topology_manager.go:138] "Creating topology manager with none policy" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.001319 4741 container_manager_linux.go:303] "Creating device plugin manager" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.002232 4741 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.002258 4741 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.002415 4741 state_mem.go:36] "Initialized new in-memory state store" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.002738 4741 server.go:1245] "Using root directory" path="/var/lib/kubelet" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.010027 4741 kubelet.go:418] "Attempting to sync node with API server" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.010050 4741 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.010076 4741 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.010088 4741 kubelet.go:324] "Adding apiserver pod source" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.010098 4741 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.013773 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.013782 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.013881 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.013877 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.015040 4741 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.015687 4741 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.017425 4741 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.019938 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.019960 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.019966 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.019972 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.019982 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.019988 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.019994 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.020004 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.020012 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.020019 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.020034 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.020040 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.022454 4741 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.022864 4741 server.go:1280] "Started kubelet" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.024063 4741 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:19 crc systemd[1]: Started Kubernetes Kubelet. Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.024394 4741 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.024437 4741 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.024954 4741 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.025819 4741 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.025858 4741 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.025909 4741 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 04:17:25.44452995 +0000 UTC Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.025949 4741 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1065h8m6.418583233s for next certificate rotation Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.025980 4741 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.026016 4741 volume_manager.go:287] "The desired_state_of_world populator starts" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.026026 4741 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.026272 4741 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.026959 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.027014 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.027429 4741 factory.go:55] Registering systemd factory Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.027498 4741 factory.go:221] Registration of the systemd container factory successfully Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.027845 4741 factory.go:153] Registering CRI-O factory Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.027861 4741 factory.go:221] Registration of the crio container factory successfully Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.028073 4741 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.028128 4741 factory.go:103] Registering Raw factory Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.028152 4741 manager.go:1196] Started watching for new ooms in manager Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.032151 4741 manager.go:319] Starting recovery of all containers Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.033755 4741 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="200ms" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.034600 4741 server.go:460] "Adding debug handlers to kubelet server" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052587 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052712 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052729 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052743 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052756 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052769 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052781 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052793 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052807 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052819 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052830 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052844 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052856 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052870 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052899 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052910 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052921 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052934 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052948 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052961 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052973 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052986 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.052997 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053009 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053023 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053035 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053050 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053064 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053076 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053089 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053158 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053171 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053184 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053198 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053211 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053587 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053611 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053625 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053636 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053648 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053661 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053673 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053687 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053700 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053712 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053726 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053738 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053750 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053764 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053777 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053790 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053801 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053819 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053835 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053849 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053862 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053876 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053923 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053937 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053950 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.053984 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054010 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054024 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054036 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054049 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054060 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054073 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054085 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054098 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054111 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054124 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054139 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054154 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054167 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054181 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054194 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054216 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054231 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054243 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054258 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054283 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054297 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054307 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054318 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054328 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054338 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054348 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054357 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054368 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054377 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054392 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054426 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054439 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054452 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054476 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054487 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054496 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054507 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054516 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054526 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054537 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054547 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054556 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054566 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054581 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054591 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054601 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054612 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054620 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054629 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054640 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054649 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054658 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054667 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054676 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054684 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054692 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054701 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054710 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054718 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054729 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054737 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054746 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054755 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054763 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054771 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054780 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054789 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054797 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054806 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054814 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054822 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054833 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054848 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054856 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054865 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.052043 4741 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.227:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1869d67c59074630 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-29 19:09:19.022827056 +0000 UTC m=+0.670616378,LastTimestamp:2025-09-29 19:09:19.022827056 +0000 UTC m=+0.670616378,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054875 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054898 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054907 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054914 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054924 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054932 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054941 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054950 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054961 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054969 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054978 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054986 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.054995 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055004 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055012 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055020 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055029 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055037 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055046 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055054 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055061 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055070 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055083 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055091 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055100 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055108 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055117 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055126 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055167 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055177 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055187 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055196 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055205 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055213 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.055223 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058782 4741 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058815 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058829 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058842 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058853 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058864 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058874 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058885 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058899 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058910 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058920 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058931 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058941 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058951 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058965 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058975 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.058984 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059000 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059010 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059020 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059032 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059045 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059055 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059066 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059078 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059089 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059099 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059109 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059119 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059129 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059139 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059150 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059163 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059176 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059188 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059199 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059209 4741 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059218 4741 reconstruct.go:97] "Volume reconstruction finished" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.059225 4741 reconciler.go:26] "Reconciler: start to sync state" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.066255 4741 manager.go:324] Recovery completed Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.075139 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.078612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.078668 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.078680 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.079827 4741 cpu_manager.go:225] "Starting CPU manager" policy="none" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.079855 4741 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.079877 4741 state_mem.go:36] "Initialized new in-memory state store" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.083109 4741 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.084597 4741 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.084635 4741 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.084662 4741 kubelet.go:2335] "Starting kubelet main sync loop" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.084700 4741 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.085153 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.085200 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.126163 4741 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.164446 4741 policy_none.go:49] "None policy: Start" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.165706 4741 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.165742 4741 state_mem.go:35] "Initializing new in-memory state store" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.185183 4741 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.221240 4741 manager.go:334] "Starting Device Plugin manager" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.221287 4741 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.221299 4741 server.go:79] "Starting device plugin registration server" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.221772 4741 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.221790 4741 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.221941 4741 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.222024 4741 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.222041 4741 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.228947 4741 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.234666 4741 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="400ms" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.322091 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.322875 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.322918 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.322928 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.322946 4741 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.323431 4741 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.385550 4741 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.385648 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.387373 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.387424 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.387435 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.387615 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.387890 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.387937 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.388284 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.388315 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.388333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.388460 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.388580 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.388654 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389088 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389110 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389118 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389187 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389219 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389244 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389259 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389261 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389299 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389245 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.389359 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390077 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390100 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390107 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390141 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390155 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390165 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390267 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390415 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.390442 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.391764 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.391844 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.391854 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.391827 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.391939 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.391945 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.392055 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.392076 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.392605 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.392624 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.392632 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463379 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463475 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463572 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463589 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463613 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463681 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463696 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463711 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463728 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463776 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463792 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463806 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463887 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463911 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.463957 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.523651 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.524980 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.525011 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.525053 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.525080 4741 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.525695 4741 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565369 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565429 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565455 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565488 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565509 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565529 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565547 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565552 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565573 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565597 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565619 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565624 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565637 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565662 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565662 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565694 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565701 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565717 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565737 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565740 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565744 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565760 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565800 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565792 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565775 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565815 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565861 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565792 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565627 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.565716 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.635231 4741 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="800ms" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.727295 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.734682 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.766113 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.774425 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.778669 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.781775 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-451afb053458f66629a3d4c5aee7590c08cec0fdda749754d14a4b3b47414fdd WatchSource:0}: Error finding container 451afb053458f66629a3d4c5aee7590c08cec0fdda749754d14a4b3b47414fdd: Status 404 returned error can't find the container with id 451afb053458f66629a3d4c5aee7590c08cec0fdda749754d14a4b3b47414fdd Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.797681 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-077c1822a230d191ef2c43d7ae25b634be9b3eed29504d1f56cfe3d44d8b3a12 WatchSource:0}: Error finding container 077c1822a230d191ef2c43d7ae25b634be9b3eed29504d1f56cfe3d44d8b3a12: Status 404 returned error can't find the container with id 077c1822a230d191ef2c43d7ae25b634be9b3eed29504d1f56cfe3d44d8b3a12 Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.798921 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6f2a550322b1980c3216a2952ae491e6bff1092052d17ca674e01e7989bf4cd9 WatchSource:0}: Error finding container 6f2a550322b1980c3216a2952ae491e6bff1092052d17ca674e01e7989bf4cd9: Status 404 returned error can't find the container with id 6f2a550322b1980c3216a2952ae491e6bff1092052d17ca674e01e7989bf4cd9 Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.800870 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-87bb8fb4aa651bdae9618e34924da0b1be1200207582b811b35f926b8b0ec700 WatchSource:0}: Error finding container 87bb8fb4aa651bdae9618e34924da0b1be1200207582b811b35f926b8b0ec700: Status 404 returned error can't find the container with id 87bb8fb4aa651bdae9618e34924da0b1be1200207582b811b35f926b8b0ec700 Sep 29 19:09:19 crc kubenswrapper[4741]: W0929 19:09:19.848118 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.848203 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.926652 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.927960 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.927991 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.928000 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:19 crc kubenswrapper[4741]: I0929 19:09:19.928022 4741 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 19:09:19 crc kubenswrapper[4741]: E0929 19:09:19.928468 4741 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.025080 4741 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.089418 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d1ced7fbe673e9f905d911f998a7840d97651a242e6227faf7f75dc91932d0e9"} Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.090493 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"87bb8fb4aa651bdae9618e34924da0b1be1200207582b811b35f926b8b0ec700"} Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.091277 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6f2a550322b1980c3216a2952ae491e6bff1092052d17ca674e01e7989bf4cd9"} Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.092934 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"077c1822a230d191ef2c43d7ae25b634be9b3eed29504d1f56cfe3d44d8b3a12"} Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.093913 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"451afb053458f66629a3d4c5aee7590c08cec0fdda749754d14a4b3b47414fdd"} Sep 29 19:09:20 crc kubenswrapper[4741]: W0929 19:09:20.430212 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:20 crc kubenswrapper[4741]: E0929 19:09:20.430613 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:20 crc kubenswrapper[4741]: E0929 19:09:20.436326 4741 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="1.6s" Sep 29 19:09:20 crc kubenswrapper[4741]: W0929 19:09:20.531330 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:20 crc kubenswrapper[4741]: E0929 19:09:20.531501 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:20 crc kubenswrapper[4741]: W0929 19:09:20.595701 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:20 crc kubenswrapper[4741]: E0929 19:09:20.595812 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.729288 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.730950 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.731009 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.731022 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:20 crc kubenswrapper[4741]: I0929 19:09:20.731064 4741 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 19:09:20 crc kubenswrapper[4741]: E0929 19:09:20.731797 4741 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.024935 4741 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.098268 4741 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="34f84a30634f0e985236321731e959a593236609ce6ca33acdb94a2aab138018" exitCode=0 Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.098368 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.098368 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"34f84a30634f0e985236321731e959a593236609ce6ca33acdb94a2aab138018"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.099340 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.099371 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.099383 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.100860 4741 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3" exitCode=0 Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.100924 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.101040 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.102620 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.102660 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.102673 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.103467 4741 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173" exitCode=0 Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.103523 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.103579 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.104726 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.104767 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.104783 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.107385 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.107436 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.107491 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.107512 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.107533 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.108458 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.108509 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.108532 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.109414 4741 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05" exitCode=0 Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.109461 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05"} Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.109536 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.110471 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.110499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.110508 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.116559 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.117490 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.117523 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:21 crc kubenswrapper[4741]: I0929 19:09:21.117532 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.025416 4741 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:22 crc kubenswrapper[4741]: E0929 19:09:22.037179 4741 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="3.2s" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.119772 4741 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b" exitCode=0 Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.119828 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.119903 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.120737 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.120765 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.120774 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.123435 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.123478 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.123488 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.123452 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.124240 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.124273 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.124285 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.125614 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.125648 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.125659 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.125668 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.125676 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.125721 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.126625 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.126660 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.126672 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.127115 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.127123 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fef753504cb83d924ff41cdcd47083961d35ebf43fe4be7293ffcccc24c09592"} Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.127140 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.127920 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.127944 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.127952 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.128089 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.128107 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.128116 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.341967 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.343211 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.343243 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.343253 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:22 crc kubenswrapper[4741]: I0929 19:09:22.343295 4741 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 19:09:22 crc kubenswrapper[4741]: E0929 19:09:22.343865 4741 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Sep 29 19:09:22 crc kubenswrapper[4741]: W0929 19:09:22.360499 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:22 crc kubenswrapper[4741]: E0929 19:09:22.360593 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:22 crc kubenswrapper[4741]: W0929 19:09:22.435762 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:22 crc kubenswrapper[4741]: E0929 19:09:22.435837 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:22 crc kubenswrapper[4741]: W0929 19:09:22.451225 4741 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Sep 29 19:09:22 crc kubenswrapper[4741]: E0929 19:09:22.451328 4741 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.131203 4741 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93" exitCode=0 Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.131738 4741 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.131464 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.131896 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.131812 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.131496 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93"} Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.131802 4741 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.132228 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133273 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133374 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133451 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133447 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133540 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133563 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133482 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133603 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133613 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133751 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133788 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.133805 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.491172 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.522985 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.523137 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.524067 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.524097 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.524107 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:23 crc kubenswrapper[4741]: I0929 19:09:23.768561 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.137863 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.137885 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.137863 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4"} Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.137966 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3"} Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.137986 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006"} Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.138002 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69"} Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.138016 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df"} Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.138882 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.138908 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.138921 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.138952 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.139001 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.139013 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:24 crc kubenswrapper[4741]: I0929 19:09:24.227850 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.141589 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.141693 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.142501 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.142527 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.142536 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.142623 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.142656 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.142666 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.544637 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.545975 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.546010 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.546021 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.546043 4741 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.560559 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.560719 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.562244 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.562330 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:25 crc kubenswrapper[4741]: I0929 19:09:25.562348 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:26 crc kubenswrapper[4741]: I0929 19:09:26.143814 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:26 crc kubenswrapper[4741]: I0929 19:09:26.146652 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:26 crc kubenswrapper[4741]: I0929 19:09:26.146690 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:26 crc kubenswrapper[4741]: I0929 19:09:26.146705 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:27 crc kubenswrapper[4741]: I0929 19:09:27.111285 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Sep 29 19:09:27 crc kubenswrapper[4741]: I0929 19:09:27.111454 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:27 crc kubenswrapper[4741]: I0929 19:09:27.112493 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:27 crc kubenswrapper[4741]: I0929 19:09:27.112534 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:27 crc kubenswrapper[4741]: I0929 19:09:27.112545 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:29 crc kubenswrapper[4741]: E0929 19:09:29.229060 4741 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.572468 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.572665 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.573901 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.573929 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.573938 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.713912 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.714069 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.715088 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.715117 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.715127 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:29 crc kubenswrapper[4741]: I0929 19:09:29.719504 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:30 crc kubenswrapper[4741]: I0929 19:09:30.150955 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:30 crc kubenswrapper[4741]: I0929 19:09:30.152516 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:30 crc kubenswrapper[4741]: I0929 19:09:30.152554 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:30 crc kubenswrapper[4741]: I0929 19:09:30.152565 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:30 crc kubenswrapper[4741]: I0929 19:09:30.154935 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.032966 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.153682 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.154631 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.154667 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.154679 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.919492 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.919708 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.921439 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.921485 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:31 crc kubenswrapper[4741]: I0929 19:09:31.921497 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.156077 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.157058 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.157088 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.157100 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.936840 4741 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.936905 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.956521 4741 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 29 19:09:32 crc kubenswrapper[4741]: I0929 19:09:32.956592 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 29 19:09:33 crc kubenswrapper[4741]: I0929 19:09:33.500667 4741 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]log ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]etcd ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/openshift.io-api-request-count-filter ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/openshift.io-startkubeinformers ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-apiserver-admission-initializer ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/generic-apiserver-start-informers ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/priority-and-fairness-config-consumer ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/priority-and-fairness-filter ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/storage-object-count-tracker-hook ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-apiextensions-informers ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-apiextensions-controllers ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/crd-informer-synced ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-system-namespaces-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-cluster-authentication-info-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-legacy-token-tracking-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-service-ip-repair-controllers ok Sep 29 19:09:33 crc kubenswrapper[4741]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Sep 29 19:09:33 crc kubenswrapper[4741]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/priority-and-fairness-config-producer ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/bootstrap-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/start-kube-aggregator-informers ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/apiservice-status-local-available-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/apiservice-status-remote-available-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/apiservice-registration-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/apiservice-wait-for-first-sync ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/apiservice-discovery-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/kube-apiserver-autoregistration ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]autoregister-completion ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/apiservice-openapi-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: [+]poststarthook/apiservice-openapiv3-controller ok Sep 29 19:09:33 crc kubenswrapper[4741]: livez check failed Sep 29 19:09:33 crc kubenswrapper[4741]: I0929 19:09:33.500727 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:09:34 crc kubenswrapper[4741]: I0929 19:09:34.033562 4741 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 29 19:09:34 crc kubenswrapper[4741]: I0929 19:09:34.033632 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 19:09:37 crc kubenswrapper[4741]: E0929 19:09:37.945964 4741 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.946874 4741 trace.go:236] Trace[1172766558]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 19:09:23.437) (total time: 14509ms): Sep 29 19:09:37 crc kubenswrapper[4741]: Trace[1172766558]: ---"Objects listed" error: 14509ms (19:09:37.946) Sep 29 19:09:37 crc kubenswrapper[4741]: Trace[1172766558]: [14.509451234s] [14.509451234s] END Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.946908 4741 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Sep 29 19:09:37 crc kubenswrapper[4741]: E0929 19:09:37.949765 4741 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.950438 4741 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.950484 4741 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.950714 4741 trace.go:236] Trace[582468243]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 19:09:26.040) (total time: 11910ms): Sep 29 19:09:37 crc kubenswrapper[4741]: Trace[582468243]: ---"Objects listed" error: 11910ms (19:09:37.950) Sep 29 19:09:37 crc kubenswrapper[4741]: Trace[582468243]: [11.910341262s] [11.910341262s] END Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.950865 4741 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.950972 4741 trace.go:236] Trace[1275038003]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 19:09:27.524) (total time: 10426ms): Sep 29 19:09:37 crc kubenswrapper[4741]: Trace[1275038003]: ---"Objects listed" error: 10426ms (19:09:37.950) Sep 29 19:09:37 crc kubenswrapper[4741]: Trace[1275038003]: [10.426170382s] [10.426170382s] END Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.950997 4741 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.971201 4741 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43280->192.168.126.11:17697: read: connection reset by peer" start-of-body= Sep 29 19:09:37 crc kubenswrapper[4741]: I0929 19:09:37.971263 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43280->192.168.126.11:17697: read: connection reset by peer" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.019125 4741 apiserver.go:52] "Watching apiserver" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.021487 4741 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.021739 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-nz2l9","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.022079 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.022154 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.022260 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.022267 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.022292 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.022352 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.024136 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.025934 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.025967 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.025968 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.026146 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.026020 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.026276 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.026324 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.026667 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.026884 4741 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.028479 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.028538 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.029122 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.029422 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.029755 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.029947 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.030098 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051359 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051442 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051466 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051485 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051505 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051526 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051544 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051593 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051613 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051637 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051656 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051678 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051695 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051712 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051731 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051754 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051774 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051800 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051826 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051847 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051834 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051867 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051892 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051936 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.051961 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052054 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052078 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052100 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052122 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052164 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052187 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052216 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052257 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052287 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052307 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052328 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052351 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052381 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052433 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052468 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052528 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052546 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052572 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052587 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052602 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052617 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052632 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052646 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052660 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052680 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052695 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052709 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052725 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052654 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.052754 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:38.552732387 +0000 UTC m=+20.200521799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057519 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057609 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057633 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057656 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057663 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057675 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057699 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057718 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057736 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057755 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057773 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057792 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057812 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057832 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057852 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057869 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057889 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057907 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057925 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057944 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057966 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057985 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058007 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058025 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058042 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058060 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058078 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058097 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058116 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058137 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058156 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058176 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058194 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058212 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058234 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058253 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058271 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058290 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058311 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058331 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058352 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058369 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058405 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058425 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058446 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058472 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058490 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058510 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058569 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.062966 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.062997 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063021 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063056 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063084 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063110 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063138 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063167 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063190 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063214 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063237 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063259 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063281 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063304 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063326 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063349 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063370 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063460 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063485 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063508 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063533 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063556 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063579 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063601 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063624 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063650 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063675 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063698 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063720 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063742 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063768 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063789 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063809 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063830 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063850 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063875 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063899 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063922 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063947 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063976 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063999 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064023 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064048 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064073 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064097 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064119 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064144 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064169 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064196 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057669 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.052931 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053091 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053242 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053455 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053597 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053717 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053749 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053837 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053941 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.053961 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.054081 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.054383 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.054446 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.054613 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.054728 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.054847 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.054862 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.055174 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.055154 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.055322 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.055702 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056037 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056240 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056364 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056488 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056608 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056646 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056663 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056666 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056784 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.056880 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057031 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057208 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057378 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057441 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057873 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057893 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.057968 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058088 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058220 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058262 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058428 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058626 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058632 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058635 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058718 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058828 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058836 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058853 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.058989 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059113 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059248 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059301 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059364 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059284 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059461 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059553 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059744 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059748 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.059896 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.060178 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.060184 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.060238 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.060467 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.060486 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.060776 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.061315 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.061379 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.061536 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.062643 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.062664 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.062905 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063084 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063103 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063125 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063350 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063452 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063602 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063623 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064639 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063778 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063859 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.063917 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064027 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064199 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064272 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064877 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.065291 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.065329 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.065407 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.065781 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.065840 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.066715 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.066756 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.067053 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.067587 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.067770 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.067934 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068060 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068155 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068350 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.064219 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068451 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068485 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068565 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068606 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068632 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068640 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068659 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068690 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068737 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068768 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068791 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068814 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068837 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068855 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068877 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068902 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068926 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068947 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068972 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.068996 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069020 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069042 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069063 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069106 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069124 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069143 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069162 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069181 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069200 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069215 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069230 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069247 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069263 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069281 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069297 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069314 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069330 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069346 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069363 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069379 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069431 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069453 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069475 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069495 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069511 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069528 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069546 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069600 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069627 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069648 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069666 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069685 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069704 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069724 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069741 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069758 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/27362580-9e01-4eb6-be43-1372e7131232-hosts-file\") pod \"node-resolver-nz2l9\" (UID: \"27362580-9e01-4eb6-be43-1372e7131232\") " pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069777 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069795 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069813 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069830 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069849 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069867 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069884 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbz86\" (UniqueName: \"kubernetes.io/projected/27362580-9e01-4eb6-be43-1372e7131232-kube-api-access-wbz86\") pod \"node-resolver-nz2l9\" (UID: \"27362580-9e01-4eb6-be43-1372e7131232\") " pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069937 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069948 4741 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069960 4741 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069969 4741 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069979 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069988 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.069999 4741 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070009 4741 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070019 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070028 4741 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070039 4741 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070047 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070056 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070066 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070075 4741 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070085 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070094 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070103 4741 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070112 4741 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070124 4741 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070134 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070143 4741 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070154 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070165 4741 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070177 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070190 4741 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070203 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070216 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070227 4741 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070554 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.070801 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071245 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071232 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071362 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071486 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071684 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071724 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071844 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071954 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.071997 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072093 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072251 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072263 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072412 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072559 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072567 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072593 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072653 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072712 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072869 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072916 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.072952 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.073181 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.073246 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.073467 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.073535 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.073774 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.073847 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:38.573826744 +0000 UTC m=+20.221616157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074003 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074096 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074116 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074338 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074648 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074798 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074808 4741 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074931 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.074953 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.075158 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.075162 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.075538 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.075599 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:38.575580831 +0000 UTC m=+20.223370233 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.075919 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.076091 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.076289 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.076328 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.076845 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.079109 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.079454 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.079541 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.079975 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.080027 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.080287 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.080299 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.080451 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.080744 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.080993 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.081141 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.081710 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.081963 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.081970 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.083283 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.083728 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.083833 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.083829 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.083880 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.084029 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.097711 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.097744 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.097756 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.097812 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:38.597794624 +0000 UTC m=+20.245583956 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.097797 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.099351 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.099374 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.099437 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.099782 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.099804 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.100108 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.100122 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.100131 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.100167 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:38.600157039 +0000 UTC m=+20.247946361 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.100641 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.100921 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.101205 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.101470 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.101624 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.101631 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.101766 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.102816 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.103016 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.103245 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.103317 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.103836 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.105064 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.105571 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.106354 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.119837 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.120166 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.120175 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.120340 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.120710 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.125695 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.126185 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.128836 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.131157 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.137450 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.138491 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.141661 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.146376 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.155086 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.156117 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.161810 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.165923 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170626 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170697 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170706 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/27362580-9e01-4eb6-be43-1372e7131232-hosts-file\") pod \"node-resolver-nz2l9\" (UID: \"27362580-9e01-4eb6-be43-1372e7131232\") " pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170755 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170816 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbz86\" (UniqueName: \"kubernetes.io/projected/27362580-9e01-4eb6-be43-1372e7131232-kube-api-access-wbz86\") pod \"node-resolver-nz2l9\" (UID: \"27362580-9e01-4eb6-be43-1372e7131232\") " pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170843 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/27362580-9e01-4eb6-be43-1372e7131232-hosts-file\") pod \"node-resolver-nz2l9\" (UID: \"27362580-9e01-4eb6-be43-1372e7131232\") " pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170857 4741 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170867 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170876 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170886 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170895 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170904 4741 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170912 4741 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170920 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170928 4741 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170937 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170944 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170952 4741 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170960 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170970 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170979 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170987 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.170995 4741 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171002 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171010 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171018 4741 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171026 4741 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171035 4741 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171043 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171051 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171058 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171067 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171086 4741 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171103 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171117 4741 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171129 4741 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171133 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171141 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171201 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171212 4741 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171222 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171231 4741 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171241 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171251 4741 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171260 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171270 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171280 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171288 4741 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171297 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171306 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171315 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171324 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171326 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171334 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171344 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171354 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171362 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171371 4741 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171381 4741 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171406 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171415 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171457 4741 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171468 4741 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171476 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171484 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171511 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171520 4741 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171528 4741 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171537 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171547 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171555 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171564 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171573 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171582 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171593 4741 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171604 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171693 4741 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171714 4741 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171726 4741 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171734 4741 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171742 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171753 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171764 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171776 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171788 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171799 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171810 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171821 4741 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171831 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171841 4741 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171850 4741 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171858 4741 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171866 4741 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171874 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171882 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171890 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171899 4741 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171932 4741 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171944 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171956 4741 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171968 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171979 4741 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.171990 4741 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172004 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172015 4741 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172026 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172034 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172043 4741 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172051 4741 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172060 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172068 4741 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172077 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172085 4741 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172093 4741 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172101 4741 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172109 4741 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172117 4741 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172124 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172132 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172140 4741 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172149 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172157 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172165 4741 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172174 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172183 4741 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172193 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172203 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172210 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172219 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172226 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172234 4741 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172243 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172250 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172257 4741 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172265 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172274 4741 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172281 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172289 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172297 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172305 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172313 4741 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172322 4741 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172330 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172338 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172346 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172353 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172360 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172367 4741 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172375 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172383 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172410 4741 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172421 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172430 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172438 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172446 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172454 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172463 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172471 4741 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172479 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172486 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172495 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172504 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172512 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172520 4741 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172529 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172536 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172544 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172553 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172560 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172568 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172577 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172586 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172594 4741 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172940 4741 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158" exitCode=255 Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.172990 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158"} Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.178947 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.192038 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbz86\" (UniqueName: \"kubernetes.io/projected/27362580-9e01-4eb6-be43-1372e7131232-kube-api-access-wbz86\") pod \"node-resolver-nz2l9\" (UID: \"27362580-9e01-4eb6-be43-1372e7131232\") " pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.192537 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.206241 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.225577 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.238557 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.255420 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.263644 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.272009 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.279771 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.285635 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.314782 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.314800 4741 scope.go:117] "RemoveContainer" containerID="ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.339046 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.350259 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.374009 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.381111 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nz2l9" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.473863 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rtxqk"] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.475376 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-ql4kz"] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.475724 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-q8tbb"] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.476783 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.477205 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.477247 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.480425 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.480632 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.480886 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.481019 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.481149 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.483878 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.484767 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.484934 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.485078 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.485191 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.485357 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.485501 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.495587 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.495678 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.505325 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.523473 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.533675 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.550499 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.557131 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.573655 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576041 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576128 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-os-release\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.576160 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:39.576142728 +0000 UTC m=+21.223932060 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576182 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-system-cni-dir\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576207 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-socket-dir-parent\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576225 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-mcd-auth-proxy-config\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576238 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-cnibin\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576253 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-netns\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576267 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576285 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-proxy-tls\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576298 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlt5c\" (UniqueName: \"kubernetes.io/projected/565a367b-9a7a-4b93-b6f8-638b639fef9b-kube-api-access-wlt5c\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576332 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-cni-bin\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576348 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-cni-multus\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576375 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-k8s-cni-cncf-io\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576439 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cni-binary-copy\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576488 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sz48\" (UniqueName: \"kubernetes.io/projected/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-kube-api-access-8sz48\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576533 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cnibin\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576558 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-os-release\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576610 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-rootfs\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576633 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-cni-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576658 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-kubelet\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576678 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-conf-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576748 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-system-cni-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.576782 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.576840 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.576909 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:39.576894332 +0000 UTC m=+21.224683674 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577119 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-etc-kubernetes\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577160 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-daemon-config\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577199 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577235 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/565a367b-9a7a-4b93-b6f8-638b639fef9b-cni-binary-copy\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577256 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-hostroot\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577278 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-multus-certs\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577308 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsl74\" (UniqueName: \"kubernetes.io/projected/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-kube-api-access-hsl74\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.577333 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-tuning-conf-dir\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.577452 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.577514 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:39.577504521 +0000 UTC m=+21.225293863 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.585120 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.599287 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.611992 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.623274 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.634825 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.642484 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.651676 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.658743 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.668259 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678167 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-mcd-auth-proxy-config\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678200 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-cnibin\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678218 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678234 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-netns\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678249 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlt5c\" (UniqueName: \"kubernetes.io/projected/565a367b-9a7a-4b93-b6f8-638b639fef9b-kube-api-access-wlt5c\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678265 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678279 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-proxy-tls\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678295 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-cni-multus\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678311 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-k8s-cni-cncf-io\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678324 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-cni-bin\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678338 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sz48\" (UniqueName: \"kubernetes.io/projected/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-kube-api-access-8sz48\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678352 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cnibin\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678366 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cni-binary-copy\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678384 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-rootfs\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678453 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-k8s-cni-cncf-io\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678468 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cnibin\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.678478 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678502 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-cni-bin\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.678512 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.678525 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.678573 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:39.678557022 +0000 UTC m=+21.326346354 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678722 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-rootfs\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678758 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-netns\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678792 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-cnibin\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678505 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-cni-multus\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679012 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-mcd-auth-proxy-config\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679070 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-os-release\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679147 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cni-binary-copy\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679246 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.678421 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-os-release\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679299 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-kubelet\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679318 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-conf-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679358 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-system-cni-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679408 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-cni-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679464 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-etc-kubernetes\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679503 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-daemon-config\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679507 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-var-lib-kubelet\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679525 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/565a367b-9a7a-4b93-b6f8-638b639fef9b-cni-binary-copy\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679577 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-hostroot\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679597 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-multus-certs\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.679627 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.679644 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679647 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsl74\" (UniqueName: \"kubernetes.io/projected/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-kube-api-access-hsl74\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.679657 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679672 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-tuning-conf-dir\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: E0929 19:09:38.679699 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:39.679684589 +0000 UTC m=+21.327474001 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679717 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-os-release\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679744 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-system-cni-dir\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679769 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-socket-dir-parent\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679880 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-socket-dir-parent\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679888 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-etc-kubernetes\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679450 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-conf-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.679928 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-system-cni-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680015 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-cni-dir\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680076 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-os-release\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680109 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-system-cni-dir\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680473 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/565a367b-9a7a-4b93-b6f8-638b639fef9b-multus-daemon-config\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680382 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680626 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-host-run-multus-certs\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680663 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/565a367b-9a7a-4b93-b6f8-638b639fef9b-hostroot\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.680984 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/565a367b-9a7a-4b93-b6f8-638b639fef9b-cni-binary-copy\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.681081 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-tuning-conf-dir\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.682794 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-proxy-tls\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.692965 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.696701 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsl74\" (UniqueName: \"kubernetes.io/projected/b2f20aa2-390a-494e-aed0-0dfd0e031d3d-kube-api-access-hsl74\") pod \"machine-config-daemon-rtxqk\" (UID: \"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\") " pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.697006 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sz48\" (UniqueName: \"kubernetes.io/projected/13be2fb8-f24c-4a58-827a-3d0a5bf8dc72-kube-api-access-8sz48\") pod \"multus-additional-cni-plugins-q8tbb\" (UID: \"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\") " pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.699680 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlt5c\" (UniqueName: \"kubernetes.io/projected/565a367b-9a7a-4b93-b6f8-638b639fef9b-kube-api-access-wlt5c\") pod \"multus-ql4kz\" (UID: \"565a367b-9a7a-4b93-b6f8-638b639fef9b\") " pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.700287 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.712736 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.799711 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" Sep 29 19:09:38 crc kubenswrapper[4741]: W0929 19:09:38.810875 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13be2fb8_f24c_4a58_827a_3d0a5bf8dc72.slice/crio-7144f022aae5f0d8814ef1954ce0374ddf545fc1e5bce2450e8f61ba264c4332 WatchSource:0}: Error finding container 7144f022aae5f0d8814ef1954ce0374ddf545fc1e5bce2450e8f61ba264c4332: Status 404 returned error can't find the container with id 7144f022aae5f0d8814ef1954ce0374ddf545fc1e5bce2450e8f61ba264c4332 Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.841257 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mrdpd"] Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.842073 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.843593 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.844771 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.844872 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.845109 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.845155 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.845416 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.847234 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.856489 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.868658 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.877818 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.887456 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.888518 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ql4kz" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.892322 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.899115 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.908550 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.921669 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.935667 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.944479 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.955235 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.962524 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.978152 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985414 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-etc-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985452 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985467 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-ovn\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985483 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-netd\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985508 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-env-overrides\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985524 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-slash\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985539 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a31e12c-13a1-4052-b679-7115e1a3574e-ovn-node-metrics-cert\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985589 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-log-socket\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985683 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-var-lib-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985762 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhpnc\" (UniqueName: \"kubernetes.io/projected/7a31e12c-13a1-4052-b679-7115e1a3574e-kube-api-access-fhpnc\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985789 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-node-log\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985851 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-netns\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985868 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-systemd\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985953 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-bin\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.985984 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-config\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.986011 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-systemd-units\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.986033 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.986057 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-script-lib\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.986083 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-kubelet\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:38 crc kubenswrapper[4741]: I0929 19:09:38.986126 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-ovn-kubernetes\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.084892 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.085019 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.086834 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-etc-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.086883 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-ovn\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.086895 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-etc-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.086910 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-netd\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.086953 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.086978 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-ovn\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087001 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-env-overrides\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087018 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a31e12c-13a1-4052-b679-7115e1a3574e-ovn-node-metrics-cert\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087021 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087038 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-slash\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087057 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-log-socket\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087072 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-var-lib-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087090 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhpnc\" (UniqueName: \"kubernetes.io/projected/7a31e12c-13a1-4052-b679-7115e1a3574e-kube-api-access-fhpnc\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087103 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-node-log\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087134 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-netns\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087152 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-systemd\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087172 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-bin\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087166 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-log-socket\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087191 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-config\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087227 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087252 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-script-lib\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087273 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-systemd-units\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-kubelet\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087304 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-ovn-kubernetes\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087364 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-ovn-kubernetes\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087665 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-env-overrides\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.086954 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-netd\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087734 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-node-log\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087766 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-var-lib-openvswitch\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087810 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-config\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087855 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-netns\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087881 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-systemd\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087903 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-bin\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087927 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-systemd-units\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087950 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-kubelet\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087975 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-slash\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.087996 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.088148 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-script-lib\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.094229 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.095087 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.096543 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.097538 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.098709 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.100924 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.101757 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.102539 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.104421 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a31e12c-13a1-4052-b679-7115e1a3574e-ovn-node-metrics-cert\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.108117 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.109141 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.118143 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.119244 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.128084 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.129279 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.129939 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.131382 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.132141 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.132644 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.137792 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.138507 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.138965 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.140019 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.140604 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.143243 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.144901 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhpnc\" (UniqueName: \"kubernetes.io/projected/7a31e12c-13a1-4052-b679-7115e1a3574e-kube-api-access-fhpnc\") pod \"ovnkube-node-mrdpd\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.145057 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.147239 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.148287 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.152657 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.154010 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.158061 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.158614 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.159519 4741 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.159641 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.162459 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.163738 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.164195 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.165814 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.166959 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.167711 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.168750 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.169683 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.170118 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.174319 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.178578 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.179207 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.180509 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.181059 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.182025 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.183229 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.184312 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.184895 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.185531 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.186530 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.187097 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.188097 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.188991 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.189022 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.189034 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"87a69fdd5c067fa858b726359d417c36d56b9506f0130379612454cf0a213466"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.189154 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.221884 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.221913 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.241497 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.246766 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.252564 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.252608 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.266814 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"b589f5b2d633c6dbb1a3d8f73457193d36281a7da86568c33a7ab08ac84352c1"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.267752 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.269865 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerStarted","Data":"6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.269909 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerStarted","Data":"58c8b06e5bb4bb08a70300c12656ee0cf434882eb449cc63968134accb6e177a"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.280497 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ac01561006d562beb1efdf8772c31b5beb8b923a30c1306fd2b3fc26ace87711"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.286156 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.286646 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5597c706970f932804fd4875c7067f82deb0c47e7781a5c0d2caf24a7ce09277"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.294886 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.298938 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerStarted","Data":"2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.298997 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerStarted","Data":"7144f022aae5f0d8814ef1954ce0374ddf545fc1e5bce2450e8f61ba264c4332"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.300486 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nz2l9" event={"ID":"27362580-9e01-4eb6-be43-1372e7131232","Type":"ContainerStarted","Data":"b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.300539 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nz2l9" event={"ID":"27362580-9e01-4eb6-be43-1372e7131232","Type":"ContainerStarted","Data":"c5aa3094a0bdb7eb2a9c0442f9ba960dc5594264b1de97670ec2a8e675da2103"} Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.308758 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.320634 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.335631 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.346160 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.358730 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.369438 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.384294 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.398032 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.418057 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.456664 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.495776 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.537859 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.576717 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.592239 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.592355 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.592417 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:41.592373363 +0000 UTC m=+23.240162705 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.592451 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.592476 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.592498 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:41.592488197 +0000 UTC m=+23.240277529 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.592583 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.592621 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:41.592612081 +0000 UTC m=+23.240401413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.616137 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.659970 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.693775 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.693857 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.693969 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.693999 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.694014 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.694045 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.694067 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.694071 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:41.694053413 +0000 UTC m=+23.341842745 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.694082 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:39 crc kubenswrapper[4741]: E0929 19:09:39.694148 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:41.694129006 +0000 UTC m=+23.341918328 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.703027 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.740291 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.775211 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:39 crc kubenswrapper[4741]: I0929 19:09:39.825196 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.085306 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.085381 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:40 crc kubenswrapper[4741]: E0929 19:09:40.085461 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:40 crc kubenswrapper[4741]: E0929 19:09:40.085584 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.304531 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2"} Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.304577 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1"} Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.305609 4741 generic.go:334] "Generic (PLEG): container finished" podID="13be2fb8-f24c-4a58-827a-3d0a5bf8dc72" containerID="2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7" exitCode=0 Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.305641 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerDied","Data":"2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7"} Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.306978 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7" exitCode=0 Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.307037 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7"} Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.307103 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"befd95673e49e8c033b3abf7605e4fdd98f7b7816eee7b0cc17e022920e7390f"} Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.325014 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.344874 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.359590 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.369792 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.384420 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.402140 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.414741 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.431098 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.444856 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.461481 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.472824 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.495934 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.510536 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.528162 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.543713 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.556003 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.570576 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.581906 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.597804 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.618416 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.657313 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.699225 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.736768 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:40 crc kubenswrapper[4741]: I0929 19:09:40.797933 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:40Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.036641 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.040938 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.044631 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.053242 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.064629 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.074557 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.085053 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.085171 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.086637 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.098112 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.112009 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.131053 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.145866 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.159673 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.197448 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.235698 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.277356 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.311899 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0"} Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.314890 4741 generic.go:334] "Generic (PLEG): container finished" podID="13be2fb8-f24c-4a58-827a-3d0a5bf8dc72" containerID="54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3" exitCode=0 Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.314990 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerDied","Data":"54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3"} Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.315711 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.318536 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61"} Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.318578 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814"} Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.318594 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1"} Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.318610 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520"} Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.318670 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff"} Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.369951 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.403183 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.436232 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.476340 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.517110 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.554695 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.600126 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.611588 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.611666 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.611707 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.611762 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:45.611738907 +0000 UTC m=+27.259528239 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.611791 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.611842 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.611856 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:45.6118429 +0000 UTC m=+27.259632232 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.611965 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:45.611949693 +0000 UTC m=+27.259739025 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.634961 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.675974 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.713146 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.713205 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713328 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713343 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713353 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713372 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713448 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713462 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713427 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:45.713415378 +0000 UTC m=+27.361204710 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:41 crc kubenswrapper[4741]: E0929 19:09:41.713566 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:45.713537272 +0000 UTC m=+27.361326684 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.715626 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.754041 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.794937 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.835577 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.878787 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.915869 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.943605 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.957186 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:41Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.957838 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Sep 29 19:09:41 crc kubenswrapper[4741]: I0929 19:09:41.979960 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.018616 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.060868 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.085519 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.085562 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:42 crc kubenswrapper[4741]: E0929 19:09:42.085723 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:42 crc kubenswrapper[4741]: E0929 19:09:42.085957 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.100057 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.136789 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.183728 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.217222 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.261290 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.296880 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.324244 4741 generic.go:334] "Generic (PLEG): container finished" podID="13be2fb8-f24c-4a58-827a-3d0a5bf8dc72" containerID="e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11" exitCode=0 Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.324304 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerDied","Data":"e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11"} Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.329219 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be"} Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.348476 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.375608 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.418855 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.457940 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.505055 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.534576 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.575365 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.615380 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.656950 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.693986 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.735045 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.781751 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.821509 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.856769 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:42 crc kubenswrapper[4741]: I0929 19:09:42.895665 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:42Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.084976 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:43 crc kubenswrapper[4741]: E0929 19:09:43.085131 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.334738 4741 generic.go:334] "Generic (PLEG): container finished" podID="13be2fb8-f24c-4a58-827a-3d0a5bf8dc72" containerID="ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c" exitCode=0 Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.334797 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerDied","Data":"ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c"} Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.349209 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.370035 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.391618 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.404260 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.422034 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.433845 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.447304 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.457957 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.469159 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.479529 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.490336 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.508524 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.520905 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:43 crc kubenswrapper[4741]: I0929 19:09:43.535751 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:43Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.085715 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.085895 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.085731 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.086226 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.340905 4741 generic.go:334] "Generic (PLEG): container finished" podID="13be2fb8-f24c-4a58-827a-3d0a5bf8dc72" containerID="c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7" exitCode=0 Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.341135 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerDied","Data":"c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.345359 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.350188 4741 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.351739 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.351786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.351798 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.351960 4741 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.362056 4741 kubelet_node_status.go:115] "Node was previously registered" node="crc" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.362296 4741 kubelet_node_status.go:79] "Successfully registered node" node="crc" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.363784 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.363863 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.363885 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.363910 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.363928 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.372147 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.384939 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.387755 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.390813 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.390959 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.391085 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.391180 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.391270 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.402407 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.404528 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.408066 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.408283 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.408446 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.408592 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.408725 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.417153 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.421347 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.424751 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.424786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.424797 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.424813 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.424824 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.431778 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.440046 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.443698 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.443737 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.443757 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.443775 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.443787 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.445846 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.458403 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: E0929 19:09:44.458551 4741 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.460286 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.460337 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.460356 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.460366 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.460380 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.460410 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.478187 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.494329 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.510022 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.519891 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.533951 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.544952 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.555996 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:44Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.562730 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.562779 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.562795 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.562819 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.562834 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.664788 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.664829 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.664838 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.664851 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.664861 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.767376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.767641 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.767718 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.767833 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.767910 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.870606 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.870640 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.870650 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.870664 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.870674 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.973035 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.973069 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.973080 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.973099 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:44 crc kubenswrapper[4741]: I0929 19:09:44.973111 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:44Z","lastTransitionTime":"2025-09-29T19:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.075876 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.075931 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.075954 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.075977 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.075995 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.085646 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.085794 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.178710 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.178774 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.178791 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.178815 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.178831 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.281729 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.281810 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.281824 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.281841 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.281854 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.353297 4741 generic.go:334] "Generic (PLEG): container finished" podID="13be2fb8-f24c-4a58-827a-3d0a5bf8dc72" containerID="d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a" exitCode=0 Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.353341 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerDied","Data":"d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.371164 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.382773 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.384892 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.384929 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.384940 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.384958 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.384970 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.395285 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.415406 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.429254 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.442219 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.456075 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.468680 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.487957 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.487988 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.487997 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.488010 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.488020 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.490089 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.502320 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.546617 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.576695 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.591315 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.591376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.591410 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.591438 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.591451 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.594046 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.607665 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:45Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.646844 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.646962 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.647016 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.647154 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.647221 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:53.647204421 +0000 UTC m=+35.294993763 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.647293 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:53.647284874 +0000 UTC m=+35.295074206 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.647335 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.647367 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:53.647358636 +0000 UTC m=+35.295147968 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.694162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.694205 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.694218 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.694236 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.694247 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.748099 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.748256 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748407 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748464 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748485 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748560 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:53.748539342 +0000 UTC m=+35.396328694 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748572 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748608 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748632 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:45 crc kubenswrapper[4741]: E0929 19:09:45.748726 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:09:53.748698557 +0000 UTC m=+35.396487929 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.800734 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.800779 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.800792 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.800809 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.800820 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.903109 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.903152 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.903162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.903176 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:45 crc kubenswrapper[4741]: I0929 19:09:45.903184 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:45Z","lastTransitionTime":"2025-09-29T19:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.005807 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.005847 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.005856 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.005871 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.005883 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.084938 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.084994 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:46 crc kubenswrapper[4741]: E0929 19:09:46.085119 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:46 crc kubenswrapper[4741]: E0929 19:09:46.085220 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.107732 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.107766 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.107775 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.107787 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.107795 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.210073 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.210117 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.210130 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.210144 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.210157 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.312954 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.312990 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.312998 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.313010 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.313019 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.361063 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" event={"ID":"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72","Type":"ContainerStarted","Data":"2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.365189 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.365423 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.373205 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.385759 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.389060 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.399341 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.408470 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.414924 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.414963 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.414974 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.414988 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.414996 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.425785 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.437775 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.451376 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.464247 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.476887 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.486112 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.498924 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.517180 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.517513 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.517550 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.517558 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.517571 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.517584 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.530141 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.542000 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.554776 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.568361 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.588176 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.599909 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.610440 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.619831 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.619876 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.619887 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.619903 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.619913 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.620376 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.633216 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.647240 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.657990 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.676524 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.690681 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.702892 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.713357 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.722057 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.722274 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.722420 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.722496 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.722575 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.727802 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.824854 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.824940 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.824957 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.824981 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.824994 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.927834 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.927876 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.927890 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.927904 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:46 crc kubenswrapper[4741]: I0929 19:09:46.927913 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:46Z","lastTransitionTime":"2025-09-29T19:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.030561 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.030600 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.030610 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.030624 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.030661 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.085657 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:47 crc kubenswrapper[4741]: E0929 19:09:47.085776 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.133168 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.133211 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.133223 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.133240 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.133265 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.235592 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.235631 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.235656 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.235672 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.235684 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.338226 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.338277 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.338287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.338303 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.338312 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.368309 4741 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.368584 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.391858 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.403642 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.415268 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.425573 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.439896 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.440109 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.440142 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.440150 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.440165 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.440181 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.451600 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.473178 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.484203 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.505437 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.519921 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.531341 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.542770 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.542805 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.542816 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.542833 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.542842 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.542876 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.555559 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.569848 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.583739 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:47Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.644833 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.644870 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.644879 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.644894 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.644936 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.747167 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.747204 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.747214 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.747228 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.747237 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.849776 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.849819 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.849829 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.849847 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.849856 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.956700 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.956927 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.957007 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.957080 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:47 crc kubenswrapper[4741]: I0929 19:09:47.957159 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:47Z","lastTransitionTime":"2025-09-29T19:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.059132 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.059188 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.059205 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.059226 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.059241 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.085725 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:48 crc kubenswrapper[4741]: E0929 19:09:48.086143 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.085761 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:48 crc kubenswrapper[4741]: E0929 19:09:48.086567 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.161926 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.161963 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.161971 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.161985 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.161994 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.264104 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.264137 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.264147 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.264161 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.264170 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.366950 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.366985 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.366994 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.367007 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.367015 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.371004 4741 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.469614 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.469679 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.469701 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.469727 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.469743 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.571989 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.572031 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.572040 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.572053 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.572062 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.674614 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.674676 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.674694 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.674719 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.674737 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.776436 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.776484 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.776499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.776520 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.776536 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.879525 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.879574 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.879587 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.879607 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.879620 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.983167 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.983226 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.983244 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.983268 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:48 crc kubenswrapper[4741]: I0929 19:09:48.983286 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:48Z","lastTransitionTime":"2025-09-29T19:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.085023 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:49 crc kubenswrapper[4741]: E0929 19:09:49.085152 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.086523 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.086568 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.086578 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.086592 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.086601 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.106626 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.123057 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.138085 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.153732 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.170494 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.187154 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.189082 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.189115 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.189126 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.189142 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.189155 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.202260 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.229365 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.249916 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.264605 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.273413 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.284356 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.290752 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.290786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.290795 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.290809 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.290819 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.297221 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.310022 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.375684 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/0.log" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.378824 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf" exitCode=1 Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.378864 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.379455 4741 scope.go:117] "RemoveContainer" containerID="61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.390112 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.392744 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.393489 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.393520 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.393541 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.393553 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.403044 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.431207 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.445837 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.460617 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.471053 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.481976 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.496035 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.496069 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.496084 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.496104 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.496120 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.499369 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.513912 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.532272 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:48Z\\\",\\\"message\\\":\\\"opping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614205 5990 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614268 5990 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614314 5990 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614224 5990 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614672 5990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 19:09:48.614511 5990 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.615186 5990 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 19:09:48.615203 5990 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 19:09:48.615221 5990 factory.go:656] Stopping watch factory\\\\nI0929 19:09:48.615232 5990 ovnkube.go:599] Stopped ovnkube\\\\nI0929 19:09:48.615233 5990 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 19\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.545201 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.560973 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.574519 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.587588 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.598335 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.598378 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.598417 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.598437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.598449 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.704699 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.704759 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.704781 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.704811 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.704827 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.807447 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.807496 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.807505 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.807518 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.807527 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.909892 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.909942 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.909955 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.909971 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:49 crc kubenswrapper[4741]: I0929 19:09:49.909982 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:49Z","lastTransitionTime":"2025-09-29T19:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.012249 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.012304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.012320 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.012339 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.012354 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.085552 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.085692 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:50 crc kubenswrapper[4741]: E0929 19:09:50.085852 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:50 crc kubenswrapper[4741]: E0929 19:09:50.085993 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.114364 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.114418 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.114429 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.114444 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.114468 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.217066 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.217118 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.217142 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.217161 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.217171 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.319384 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.319521 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.319540 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.319564 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.319585 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.384720 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/1.log" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.385337 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/0.log" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.388886 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3" exitCode=1 Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.388951 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.389006 4741 scope.go:117] "RemoveContainer" containerID="61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.391468 4741 scope.go:117] "RemoveContainer" containerID="45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3" Sep 29 19:09:50 crc kubenswrapper[4741]: E0929 19:09:50.391783 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.404441 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.422436 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.422477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.422488 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.422503 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.422515 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.425354 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.438230 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.449325 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.460662 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.474098 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.487071 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.501777 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.516080 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.524893 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.524958 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.524977 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.525001 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.525017 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.530226 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.545530 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.562265 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.572780 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.597642 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:48Z\\\",\\\"message\\\":\\\"opping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614205 5990 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614268 5990 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614314 5990 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614224 5990 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614672 5990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 19:09:48.614511 5990 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.615186 5990 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 19:09:48.615203 5990 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 19:09:48.615221 5990 factory.go:656] Stopping watch factory\\\\nI0929 19:09:48.615232 5990 ovnkube.go:599] Stopped ovnkube\\\\nI0929 19:09:48.615233 5990 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 19\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.627424 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.627459 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.627468 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.627482 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.627534 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.637094 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m"] Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.637489 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.640586 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.640599 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.663537 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.678832 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.691888 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.703210 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.714799 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.724983 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.729091 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.729140 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.729151 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.729165 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.729175 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.737232 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.748986 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.758739 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.776452 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.792840 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/388588a6-1c5b-4478-8895-40997b0e00c6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.792896 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/388588a6-1c5b-4478-8895-40997b0e00c6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.792956 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/388588a6-1c5b-4478-8895-40997b0e00c6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.792993 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wkrj\" (UniqueName: \"kubernetes.io/projected/388588a6-1c5b-4478-8895-40997b0e00c6-kube-api-access-4wkrj\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.795870 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.818200 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.828746 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.831156 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.831184 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.831193 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.831206 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.831214 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.846232 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:48Z\\\",\\\"message\\\":\\\"opping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614205 5990 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614268 5990 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614314 5990 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614224 5990 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614672 5990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 19:09:48.614511 5990 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.615186 5990 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 19:09:48.615203 5990 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 19:09:48.615221 5990 factory.go:656] Stopping watch factory\\\\nI0929 19:09:48.615232 5990 ovnkube.go:599] Stopped ovnkube\\\\nI0929 19:09:48.615233 5990 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 19\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.857961 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.893966 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/388588a6-1c5b-4478-8895-40997b0e00c6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.894000 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/388588a6-1c5b-4478-8895-40997b0e00c6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.894025 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/388588a6-1c5b-4478-8895-40997b0e00c6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.894044 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wkrj\" (UniqueName: \"kubernetes.io/projected/388588a6-1c5b-4478-8895-40997b0e00c6-kube-api-access-4wkrj\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.894635 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/388588a6-1c5b-4478-8895-40997b0e00c6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.894767 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/388588a6-1c5b-4478-8895-40997b0e00c6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.902717 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/388588a6-1c5b-4478-8895-40997b0e00c6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.916691 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wkrj\" (UniqueName: \"kubernetes.io/projected/388588a6-1c5b-4478-8895-40997b0e00c6-kube-api-access-4wkrj\") pod \"ovnkube-control-plane-749d76644c-tlr7m\" (UID: \"388588a6-1c5b-4478-8895-40997b0e00c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.933706 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.933741 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.933750 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.933764 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.933772 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:50Z","lastTransitionTime":"2025-09-29T19:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.950333 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" Sep 29 19:09:50 crc kubenswrapper[4741]: W0929 19:09:50.960919 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod388588a6_1c5b_4478_8895_40997b0e00c6.slice/crio-24bb8c1d992b4c1b9ce2ba41b1c03a17b776b819d28da85598db9233f5a7bef0 WatchSource:0}: Error finding container 24bb8c1d992b4c1b9ce2ba41b1c03a17b776b819d28da85598db9233f5a7bef0: Status 404 returned error can't find the container with id 24bb8c1d992b4c1b9ce2ba41b1c03a17b776b819d28da85598db9233f5a7bef0 Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.997318 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-fwxc2"] Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.997656 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.999476 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.999536 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.999559 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Sep 29 19:09:50 crc kubenswrapper[4741]: I0929 19:09:50.999487 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.009680 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.021952 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.032418 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.035898 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.035943 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.035956 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.035974 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.035987 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.053052 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.067208 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.079069 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.085193 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:51 crc kubenswrapper[4741]: E0929 19:09:51.085310 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.091803 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.096140 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95b724d8-e45d-4f98-86be-85df7d4a57da-host\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.096173 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/95b724d8-e45d-4f98-86be-85df7d4a57da-serviceca\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.096194 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztjnp\" (UniqueName: \"kubernetes.io/projected/95b724d8-e45d-4f98-86be-85df7d4a57da-kube-api-access-ztjnp\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.107633 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.118764 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.130674 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.139419 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.139450 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.139459 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.139472 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.139480 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.150866 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:48Z\\\",\\\"message\\\":\\\"opping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614205 5990 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614268 5990 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614314 5990 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614224 5990 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614672 5990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 19:09:48.614511 5990 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.615186 5990 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 19:09:48.615203 5990 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 19:09:48.615221 5990 factory.go:656] Stopping watch factory\\\\nI0929 19:09:48.615232 5990 ovnkube.go:599] Stopped ovnkube\\\\nI0929 19:09:48.615233 5990 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 19\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.161666 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.175350 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.187967 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.197367 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztjnp\" (UniqueName: \"kubernetes.io/projected/95b724d8-e45d-4f98-86be-85df7d4a57da-kube-api-access-ztjnp\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.197452 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95b724d8-e45d-4f98-86be-85df7d4a57da-host\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.197469 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/95b724d8-e45d-4f98-86be-85df7d4a57da-serviceca\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.197569 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95b724d8-e45d-4f98-86be-85df7d4a57da-host\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.197976 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.198313 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/95b724d8-e45d-4f98-86be-85df7d4a57da-serviceca\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.212044 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.212319 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztjnp\" (UniqueName: \"kubernetes.io/projected/95b724d8-e45d-4f98-86be-85df7d4a57da-kube-api-access-ztjnp\") pod \"node-ca-fwxc2\" (UID: \"95b724d8-e45d-4f98-86be-85df7d4a57da\") " pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.241678 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.241710 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.241718 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.241730 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.241741 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.308549 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fwxc2" Sep 29 19:09:51 crc kubenswrapper[4741]: W0929 19:09:51.318479 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95b724d8_e45d_4f98_86be_85df7d4a57da.slice/crio-b408461537ae4c6eb64f43bf79b8cee15153b8b7fc7b36dc734418608ca3f7fd WatchSource:0}: Error finding container b408461537ae4c6eb64f43bf79b8cee15153b8b7fc7b36dc734418608ca3f7fd: Status 404 returned error can't find the container with id b408461537ae4c6eb64f43bf79b8cee15153b8b7fc7b36dc734418608ca3f7fd Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.343440 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.343473 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.343481 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.343495 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.343507 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.394791 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fwxc2" event={"ID":"95b724d8-e45d-4f98-86be-85df7d4a57da","Type":"ContainerStarted","Data":"b408461537ae4c6eb64f43bf79b8cee15153b8b7fc7b36dc734418608ca3f7fd"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.398901 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" event={"ID":"388588a6-1c5b-4478-8895-40997b0e00c6","Type":"ContainerStarted","Data":"a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.398957 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" event={"ID":"388588a6-1c5b-4478-8895-40997b0e00c6","Type":"ContainerStarted","Data":"8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.398972 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" event={"ID":"388588a6-1c5b-4478-8895-40997b0e00c6","Type":"ContainerStarted","Data":"24bb8c1d992b4c1b9ce2ba41b1c03a17b776b819d28da85598db9233f5a7bef0"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.401133 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/1.log" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.406787 4741 scope.go:117] "RemoveContainer" containerID="45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3" Sep 29 19:09:51 crc kubenswrapper[4741]: E0929 19:09:51.406920 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.411035 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.430141 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.442026 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.444898 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.444942 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.444957 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.444974 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.444986 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.453078 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.461034 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.471896 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.481750 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.492713 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.503252 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.514973 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.526299 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.536557 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.547528 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.547569 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.547578 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.547592 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.547758 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.549259 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.558814 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.575134 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61defab90fae6bc1a843a20f341d57092bf08a97c019ced6bfcb96eab26a9cbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:48Z\\\",\\\"message\\\":\\\"opping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614205 5990 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614268 5990 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614314 5990 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614224 5990 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.614672 5990 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 19:09:48.614511 5990 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 19:09:48.615186 5990 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 19:09:48.615203 5990 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 19:09:48.615221 5990 factory.go:656] Stopping watch factory\\\\nI0929 19:09:48.615232 5990 ovnkube.go:599] Stopped ovnkube\\\\nI0929 19:09:48.615233 5990 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 19\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.586233 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.595355 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.606538 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.615740 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.633744 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.645896 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.650379 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.650424 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.650433 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.650451 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.650460 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.659804 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.670110 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.683030 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.695499 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.704801 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.721035 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.730975 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.741895 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.752496 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.752546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.752558 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.752576 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.752589 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.753629 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.763773 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.775043 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.855255 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.855350 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.855360 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.855376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.855400 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.957646 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.957688 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.957699 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.957714 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:51 crc kubenswrapper[4741]: I0929 19:09:51.957723 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:51Z","lastTransitionTime":"2025-09-29T19:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.060265 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.060322 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.060336 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.060353 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.060364 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.085662 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:52 crc kubenswrapper[4741]: E0929 19:09:52.085819 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.085672 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:52 crc kubenswrapper[4741]: E0929 19:09:52.086011 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.099059 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7krvl"] Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.099941 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:52 crc kubenswrapper[4741]: E0929 19:09:52.100069 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.120045 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.139478 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.150797 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.163318 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.163366 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.163376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.163409 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.163422 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.175729 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.189167 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.200702 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.207313 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmf6f\" (UniqueName: \"kubernetes.io/projected/58436159-f44b-45ad-98f9-a60bc389bc7b-kube-api-access-jmf6f\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.207374 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.211295 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.227980 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.240064 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.248732 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.264852 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.265377 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.265415 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.265426 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.265443 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.265456 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.274069 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.281818 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.292571 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.303002 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.308185 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmf6f\" (UniqueName: \"kubernetes.io/projected/58436159-f44b-45ad-98f9-a60bc389bc7b-kube-api-access-jmf6f\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.308237 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:52 crc kubenswrapper[4741]: E0929 19:09:52.308360 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:52 crc kubenswrapper[4741]: E0929 19:09:52.308435 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:52.808418138 +0000 UTC m=+34.456207470 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.312280 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.322631 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmf6f\" (UniqueName: \"kubernetes.io/projected/58436159-f44b-45ad-98f9-a60bc389bc7b-kube-api-access-jmf6f\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.341690 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.367903 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.367954 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.367966 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.367979 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.367988 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.409687 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fwxc2" event={"ID":"95b724d8-e45d-4f98-86be-85df7d4a57da","Type":"ContainerStarted","Data":"c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.422234 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.434871 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.464637 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.470589 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.470627 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.470637 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.470653 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.470684 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.499679 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.542522 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.573477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.573536 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.573556 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.573583 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.573602 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.583107 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.628198 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.661939 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.676551 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.676617 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.676640 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.676668 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.676689 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.704719 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.740455 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.780034 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.780097 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.780116 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.780141 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.780157 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.783548 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.813285 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:52 crc kubenswrapper[4741]: E0929 19:09:52.813536 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:52 crc kubenswrapper[4741]: E0929 19:09:52.813601 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:53.813584721 +0000 UTC m=+35.461374073 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.822138 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.860313 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.882582 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.882631 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.882647 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.882669 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.882684 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.906955 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.938520 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.980812 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.985761 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.985801 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.985819 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.985839 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:52 crc kubenswrapper[4741]: I0929 19:09:52.985852 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:52Z","lastTransitionTime":"2025-09-29T19:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.021359 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.085260 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.085490 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.087761 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.087804 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.087815 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.087830 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.087842 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.189718 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.189755 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.189767 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.189781 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.189793 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.293204 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.293253 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.293267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.293286 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.293299 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.396454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.396506 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.396518 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.396534 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.396546 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.499230 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.499278 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.499289 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.499308 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.499319 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.602105 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.602477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.602511 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.602546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.602570 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.705258 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.705308 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.705317 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.705330 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.705340 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.722690 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.722819 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:10:09.722795485 +0000 UTC m=+51.370584897 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.722905 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.722969 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.723027 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.723102 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:09.723082284 +0000 UTC m=+51.370871636 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.723116 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.723155 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:09.723146716 +0000 UTC m=+51.370936158 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.774559 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.789311 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.801953 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.807664 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.807707 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.807716 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.807732 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.807741 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.824843 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.824964 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.825056 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.825366 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.825440 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.825486 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.825583 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:09.825543971 +0000 UTC m=+51.473333343 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.826450 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.826489 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.826510 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.826593 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:09.826569453 +0000 UTC m=+51.474358825 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.826730 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: E0929 19:09:53.826806 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:55.826772499 +0000 UTC m=+37.474561881 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.823267 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.841996 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.852828 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.865101 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.879251 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.890600 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.903066 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.909989 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.910019 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.910030 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.910047 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.910067 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:53Z","lastTransitionTime":"2025-09-29T19:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.922898 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.936593 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.946526 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.962922 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.973539 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.983657 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:53 crc kubenswrapper[4741]: I0929 19:09:53.991472 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.005771 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.012382 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.012426 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.012435 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.012448 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.012459 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.085339 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.085410 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.085475 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.085420 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.085565 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.085651 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.115456 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.115493 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.115506 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.115520 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.115530 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.217877 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.217911 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.217920 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.217933 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.217941 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.319968 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.320009 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.320022 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.320038 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.320049 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.421957 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.422023 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.422039 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.422064 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.422083 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.524841 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.524879 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.524891 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.524906 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.524918 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.627263 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.627297 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.627309 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.627324 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.627335 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.729586 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.729636 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.729650 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.729669 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.729683 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.758121 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.758199 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.758329 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.758442 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.758459 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.775799 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.780215 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.780285 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.780305 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.780332 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.780351 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.796425 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.801182 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.801230 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.801247 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.801267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.801281 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.819856 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.827784 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.827892 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.827926 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.827969 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.827998 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.847938 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.855358 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.856132 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.856147 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.856171 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.856186 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.870777 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:54 crc kubenswrapper[4741]: E0929 19:09:54.870909 4741 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.872972 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.873061 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.873090 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.873124 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.873149 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.976184 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.976248 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.976266 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.976291 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:54 crc kubenswrapper[4741]: I0929 19:09:54.976310 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:54Z","lastTransitionTime":"2025-09-29T19:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.079815 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.079874 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.079894 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.079919 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.079938 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.085179 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:55 crc kubenswrapper[4741]: E0929 19:09:55.085301 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.182372 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.182437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.182449 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.182465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.182477 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.286112 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.286200 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.286222 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.286248 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.286268 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.389172 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.389227 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.389240 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.389261 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.389276 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.492653 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.492729 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.492749 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.492779 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.492799 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.595769 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.595848 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.595873 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.595914 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.595943 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.699485 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.699535 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.699547 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.699564 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.699579 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.803004 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.803082 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.803100 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.803131 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.803156 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.844851 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:55 crc kubenswrapper[4741]: E0929 19:09:55.845062 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:55 crc kubenswrapper[4741]: E0929 19:09:55.845187 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:09:59.845151012 +0000 UTC m=+41.492940374 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.906452 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.906526 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.906538 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.906566 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:55 crc kubenswrapper[4741]: I0929 19:09:55.906587 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:55Z","lastTransitionTime":"2025-09-29T19:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.009952 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.010044 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.010080 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.010114 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.010138 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.085524 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.085681 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:56 crc kubenswrapper[4741]: E0929 19:09:56.085753 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.085777 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:56 crc kubenswrapper[4741]: E0929 19:09:56.085911 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:56 crc kubenswrapper[4741]: E0929 19:09:56.086081 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.112725 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.112782 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.112792 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.112810 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.112820 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.215116 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.215157 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.215166 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.215180 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.215189 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.317807 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.317837 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.317845 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.317857 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.317867 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.421615 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.421671 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.421683 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.421707 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.421721 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.524789 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.524845 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.524855 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.524882 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.524898 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.628273 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.628349 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.628367 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.628438 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.628458 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.730955 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.731015 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.731035 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.731060 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.731073 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.833479 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.833526 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.833537 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.833555 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.833566 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.935801 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.935849 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.935861 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.935875 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:56 crc kubenswrapper[4741]: I0929 19:09:56.935896 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:56Z","lastTransitionTime":"2025-09-29T19:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.038429 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.038482 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.038494 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.038510 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.038520 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.085494 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:57 crc kubenswrapper[4741]: E0929 19:09:57.085632 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.141138 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.141186 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.141199 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.141218 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.141231 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.243308 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.243343 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.243351 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.243364 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.243375 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.346178 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.346243 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.346253 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.346267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.346275 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.448782 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.448850 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.448868 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.448891 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.448945 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.551919 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.551961 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.551974 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.551990 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.552001 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.654722 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.654767 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.654780 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.654796 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.654808 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.757239 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.757277 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.757290 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.757306 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.757320 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.861096 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.861147 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.861182 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.861201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.861215 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.963622 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.963662 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.963672 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.963691 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:57 crc kubenswrapper[4741]: I0929 19:09:57.963702 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:57Z","lastTransitionTime":"2025-09-29T19:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.067084 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.067162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.067180 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.067206 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.067225 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.085809 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.085911 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:09:58 crc kubenswrapper[4741]: E0929 19:09:58.085986 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.085917 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:09:58 crc kubenswrapper[4741]: E0929 19:09:58.086310 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:09:58 crc kubenswrapper[4741]: E0929 19:09:58.086130 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.169790 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.169845 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.169858 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.169876 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.169889 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.272229 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.272291 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.272304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.272319 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.272330 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.374226 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.374315 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.374343 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.374372 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.374420 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.476773 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.476819 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.476830 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.476847 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.476860 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.579297 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.579351 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.579367 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.579498 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.579518 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.682751 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.682824 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.682842 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.682866 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.682884 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.785891 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.785947 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.785964 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.785986 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.786004 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.888298 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.888375 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.888432 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.888458 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.888475 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.991501 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.991569 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.991589 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.991612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:58 crc kubenswrapper[4741]: I0929 19:09:58.991629 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:58Z","lastTransitionTime":"2025-09-29T19:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.085988 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:09:59 crc kubenswrapper[4741]: E0929 19:09:59.086220 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.093795 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.093861 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.093879 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.093904 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.093921 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.113338 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.135354 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.155493 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.167213 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.180565 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.191192 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.195691 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.195731 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.195742 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.195759 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.195772 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.204221 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.215862 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.230663 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.245886 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.255595 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.267352 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.276804 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.292209 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.298022 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.298287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.298454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.298527 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.298588 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.304475 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.314254 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.326889 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.401348 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.401407 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.401420 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.401437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.401449 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.503564 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.503603 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.503612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.503625 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.503638 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.606201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.606243 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.606255 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.606269 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.606279 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.708443 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.708479 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.708487 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.708501 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.708510 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.811003 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.811050 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.811067 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.811085 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.811098 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.887067 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:09:59 crc kubenswrapper[4741]: E0929 19:09:59.887280 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:59 crc kubenswrapper[4741]: E0929 19:09:59.887455 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:10:07.887367819 +0000 UTC m=+49.535157181 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.913478 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.913520 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.913530 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.913546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:09:59 crc kubenswrapper[4741]: I0929 19:09:59.913558 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:09:59Z","lastTransitionTime":"2025-09-29T19:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.015728 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.015818 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.015845 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.015871 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.015888 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.085648 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.085698 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:00 crc kubenswrapper[4741]: E0929 19:10:00.085791 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.085647 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:00 crc kubenswrapper[4741]: E0929 19:10:00.085936 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:00 crc kubenswrapper[4741]: E0929 19:10:00.086045 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.118366 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.118464 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.118503 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.118538 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.118560 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.220698 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.220792 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.220810 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.220869 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.220888 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.324268 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.324329 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.324368 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.324435 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.324472 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.337581 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.339149 4741 scope.go:117] "RemoveContainer" containerID="45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.427876 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.428437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.428536 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.428617 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.428684 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.440687 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/1.log" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.531365 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.531411 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.531424 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.531441 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.531453 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.633679 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.633715 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.633728 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.633743 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.633753 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.736321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.736361 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.736376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.736403 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.736413 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.839066 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.839319 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.839426 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.839516 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.839616 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.941343 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.941697 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.941786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.941866 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:00 crc kubenswrapper[4741]: I0929 19:10:00.941949 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:00Z","lastTransitionTime":"2025-09-29T19:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.044038 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.044297 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.044481 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.044667 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.044749 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.084942 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:01 crc kubenswrapper[4741]: E0929 19:10:01.085261 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.146753 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.146786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.146794 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.146810 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.146819 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.248782 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.249070 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.249153 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.249235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.249321 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.352786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.352833 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.352845 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.352859 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.352870 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.448338 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/2.log" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.448974 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/1.log" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.452440 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3" exitCode=1 Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.452547 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.452613 4741 scope.go:117] "RemoveContainer" containerID="45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.453947 4741 scope.go:117] "RemoveContainer" containerID="3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3" Sep 29 19:10:01 crc kubenswrapper[4741]: E0929 19:10:01.454271 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.456227 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.456281 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.456292 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.456306 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.456316 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.470171 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.485456 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.517973 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.539221 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.552003 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.559267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.559307 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.559319 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.559337 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.559350 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.562972 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.576835 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.590319 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.607048 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.632867 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.645081 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.657160 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.661118 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.661158 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.661174 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.661197 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.661214 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.669198 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.680597 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.692902 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.708170 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.720085 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.764086 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.764173 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.764188 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.764207 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.764223 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.866172 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.866223 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.866231 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.866244 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.866254 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.968579 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.968615 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.968623 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.968638 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:01 crc kubenswrapper[4741]: I0929 19:10:01.968646 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:01Z","lastTransitionTime":"2025-09-29T19:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.073449 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.073497 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.073514 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.073533 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.073545 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.084965 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.084989 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:02 crc kubenswrapper[4741]: E0929 19:10:02.085810 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.085006 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:02 crc kubenswrapper[4741]: E0929 19:10:02.085898 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:02 crc kubenswrapper[4741]: E0929 19:10:02.086000 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.176225 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.176258 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.176267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.176279 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.176288 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.279286 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.279341 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.279355 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.279377 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.279422 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.381587 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.381637 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.381648 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.381667 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.381679 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.458142 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/2.log" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.483785 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.483935 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.483992 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.484039 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.484056 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.587470 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.587540 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.587567 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.587597 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.587619 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.690291 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.690330 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.690374 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.690420 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.690436 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.792874 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.792933 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.792948 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.792969 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.792984 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.895155 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.895208 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.895223 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.895245 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.895259 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.998152 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.998186 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.998195 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.998208 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:02 crc kubenswrapper[4741]: I0929 19:10:02.998217 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:02Z","lastTransitionTime":"2025-09-29T19:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.085212 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:03 crc kubenswrapper[4741]: E0929 19:10:03.085582 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.100550 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.100598 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.100612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.100667 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.100682 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.202543 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.202581 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.202594 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.202610 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.202621 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.305260 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.305304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.305314 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.305329 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.305341 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.407699 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.407739 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.407748 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.407762 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.407772 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.509712 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.509748 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.509758 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.509772 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.509780 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.611783 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.611824 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.611833 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.611848 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.611856 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.713901 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.713940 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.713948 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.713963 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.713977 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.816334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.816414 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.816432 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.816450 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.816462 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.918665 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.918712 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.918724 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.918740 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:03 crc kubenswrapper[4741]: I0929 19:10:03.918751 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:03Z","lastTransitionTime":"2025-09-29T19:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.021228 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.021294 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.021311 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.021337 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.021355 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.085076 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.085113 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.085154 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:04 crc kubenswrapper[4741]: E0929 19:10:04.085209 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:04 crc kubenswrapper[4741]: E0929 19:10:04.085279 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:04 crc kubenswrapper[4741]: E0929 19:10:04.085373 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.124163 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.124284 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.124297 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.124318 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.124330 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.226244 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.226277 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.226286 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.226323 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.226332 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.329488 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.329524 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.329535 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.329553 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.329564 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.432425 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.432482 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.432499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.432523 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.432540 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.535292 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.535331 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.535339 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.535354 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.535364 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.637973 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.638006 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.638017 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.638032 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.638045 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.740151 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.740415 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.740425 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.740441 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.740450 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.843228 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.843273 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.843287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.843304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.843317 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.945904 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.945946 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.945956 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.945972 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:04 crc kubenswrapper[4741]: I0929 19:10:04.945983 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:04Z","lastTransitionTime":"2025-09-29T19:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.028647 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.028681 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.028691 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.028705 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.028713 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: E0929 19:10:05.040919 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:05Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.044144 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.044181 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.044191 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.044205 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.044215 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: E0929 19:10:05.056609 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:05Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.060317 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.060377 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.060417 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.060451 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.060468 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: E0929 19:10:05.076516 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:05Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.080437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.080474 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.080483 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.080499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.080508 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.085792 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:05 crc kubenswrapper[4741]: E0929 19:10:05.086115 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:05 crc kubenswrapper[4741]: E0929 19:10:05.093244 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:05Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.097923 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.097953 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.097961 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.097973 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.097982 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: E0929 19:10:05.110014 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:05Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:05 crc kubenswrapper[4741]: E0929 19:10:05.110156 4741 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.111739 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.111791 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.111807 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.111829 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.111847 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.213710 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.213747 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.213756 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.213789 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.213799 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.317869 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.317899 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.317909 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.317921 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.317932 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.421154 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.421201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.421213 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.421231 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.421243 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.526000 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.526067 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.526091 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.526122 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.526145 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.629121 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.629156 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.629165 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.629177 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.629187 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.731902 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.731942 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.731953 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.731970 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.731981 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.834247 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.834287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.834297 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.834311 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.834321 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.936444 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.936491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.936507 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.936528 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:05 crc kubenswrapper[4741]: I0929 19:10:05.936543 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:05Z","lastTransitionTime":"2025-09-29T19:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.038703 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.038766 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.038781 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.038801 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.038816 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.085722 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:06 crc kubenswrapper[4741]: E0929 19:10:06.085847 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.085722 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:06 crc kubenswrapper[4741]: E0929 19:10:06.085904 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.085722 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:06 crc kubenswrapper[4741]: E0929 19:10:06.085959 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.140933 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.140983 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.141011 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.141028 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.141038 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.243454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.243519 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.243529 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.243544 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.243552 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.345954 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.345993 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.346002 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.346016 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.346025 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.448629 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.448697 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.448709 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.448724 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.448733 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.550588 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.550627 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.550638 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.550654 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.550665 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.652955 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.653017 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.653029 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.653046 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.653058 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.755890 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.755973 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.755992 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.756030 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.756066 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.858273 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.858308 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.858321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.858335 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.858344 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.960594 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.960649 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.960657 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.960670 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:06 crc kubenswrapper[4741]: I0929 19:10:06.960678 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:06Z","lastTransitionTime":"2025-09-29T19:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.062570 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.062618 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.062630 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.062649 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.062682 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.084910 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:07 crc kubenswrapper[4741]: E0929 19:10:07.085048 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.164859 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.164895 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.164903 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.164916 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.164925 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.267058 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.267115 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.267138 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.267170 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.267193 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.369420 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.369479 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.369498 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.369521 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.369538 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.471912 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.471985 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.472001 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.472055 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.472131 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.574029 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.574078 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.574090 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.574108 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.574121 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.676674 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.676708 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.676718 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.676733 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.676761 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.779823 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.780122 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.780194 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.780278 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.780344 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.882541 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.882606 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.882627 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.882651 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.882674 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.966582 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:07 crc kubenswrapper[4741]: E0929 19:10:07.966713 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:10:07 crc kubenswrapper[4741]: E0929 19:10:07.966778 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:10:23.966758836 +0000 UTC m=+65.614548168 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.985262 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.985313 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.985334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.985365 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:07 crc kubenswrapper[4741]: I0929 19:10:07.985382 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:07Z","lastTransitionTime":"2025-09-29T19:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.085450 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.085536 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:08 crc kubenswrapper[4741]: E0929 19:10:08.085568 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:08 crc kubenswrapper[4741]: E0929 19:10:08.085780 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.085462 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:08 crc kubenswrapper[4741]: E0929 19:10:08.086029 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.087159 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.087183 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.087192 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.087205 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.087215 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.189937 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.189967 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.189975 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.189991 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.190000 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.292549 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.292596 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.292606 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.292623 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.292635 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.394361 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.394408 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.394418 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.394432 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.394441 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.496600 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.496659 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.496671 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.496693 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.496704 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.599902 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.599951 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.599962 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.599979 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.599993 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.702016 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.702054 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.702063 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.702077 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.702087 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.805245 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.805297 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.805308 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.805328 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.805343 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.908112 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.908173 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.908184 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.908199 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:08 crc kubenswrapper[4741]: I0929 19:10:08.908208 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:08Z","lastTransitionTime":"2025-09-29T19:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.010309 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.010346 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.010357 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.010370 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.010379 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.085512 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.085637 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.100225 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.112546 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.113016 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.113056 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.113068 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.113086 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.113098 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.127243 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.142635 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.155189 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.168474 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.181177 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.202126 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.212209 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.214884 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.214919 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.214930 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.214945 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.214955 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.223045 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.234366 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.252928 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.263859 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.274760 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.284855 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.295763 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.304968 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.317528 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.317562 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.317573 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.317588 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.317599 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.419637 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.419670 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.419678 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.419691 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.419699 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.521974 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.522013 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.522025 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.522068 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.522081 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.577037 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.589090 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.590970 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.609853 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.622252 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.623893 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.623923 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.623931 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.623943 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.623952 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.633474 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.643484 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.655714 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.666704 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.677126 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.693292 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.706368 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.720976 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.726006 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.726041 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.726052 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.726068 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.726080 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.735490 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.746462 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.755208 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.766094 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.775578 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.786435 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.786593 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:10:41.786569058 +0000 UTC m=+83.434358410 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.786701 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.786779 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.786792 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.786846 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:41.786831335 +0000 UTC m=+83.434620667 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.786964 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.787043 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:41.787027572 +0000 UTC m=+83.434816904 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.792303 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45fee242d6e0df357f608f5982b1f89bc1c3eee8b8b99f55bde1c185bdac92e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"message\\\":\\\"c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:09:50.083961 6132 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0929 19:09:50.084037 6132 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:09Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.828801 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.828841 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.828850 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.828864 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.828874 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.887924 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.887987 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888104 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888116 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888149 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888160 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888207 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:41.888192347 +0000 UTC m=+83.535981679 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888121 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888243 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:09 crc kubenswrapper[4741]: E0929 19:10:09.888285 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:10:41.888271959 +0000 UTC m=+83.536061291 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.931071 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.931103 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.931114 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.931128 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:09 crc kubenswrapper[4741]: I0929 19:10:09.931139 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:09Z","lastTransitionTime":"2025-09-29T19:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.033491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.033533 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.033544 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.033588 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.033606 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.085423 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.085423 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:10 crc kubenswrapper[4741]: E0929 19:10:10.085578 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.085446 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:10 crc kubenswrapper[4741]: E0929 19:10:10.085703 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:10 crc kubenswrapper[4741]: E0929 19:10:10.085796 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.136064 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.136129 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.136143 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.136160 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.136173 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.238869 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.238927 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.238944 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.238967 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.238986 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.342326 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.342739 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.342933 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.343139 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.343333 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.446025 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.446058 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.446067 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.446081 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.446089 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.548856 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.549056 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.549155 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.549229 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.549309 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.653204 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.653516 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.653704 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.653825 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.653922 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.756730 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.756973 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.757038 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.757108 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.757205 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.860488 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.860533 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.860546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.860562 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.860573 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.962456 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.962733 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.962815 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.962879 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:10 crc kubenswrapper[4741]: I0929 19:10:10.962934 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:10Z","lastTransitionTime":"2025-09-29T19:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.065045 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.065557 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.065642 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.065733 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.065803 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.084912 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:11 crc kubenswrapper[4741]: E0929 19:10:11.085072 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.167872 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.167904 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.167911 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.167924 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.167932 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.270111 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.270352 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.270465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.270586 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.270683 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.376465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.376519 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.376530 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.376545 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.376562 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.479621 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.479692 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.479716 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.479746 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.479767 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.582920 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.582962 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.582972 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.582986 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.582995 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.686483 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.686533 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.686552 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.686572 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.686589 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.789098 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.789149 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.789167 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.789191 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.789210 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.891467 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.891504 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.891513 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.891527 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.891537 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.993328 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.993371 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.993379 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.993405 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:11 crc kubenswrapper[4741]: I0929 19:10:11.993415 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:11Z","lastTransitionTime":"2025-09-29T19:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.085261 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.085261 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:12 crc kubenswrapper[4741]: E0929 19:10:12.085424 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:12 crc kubenswrapper[4741]: E0929 19:10:12.085493 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.085294 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:12 crc kubenswrapper[4741]: E0929 19:10:12.085574 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.095929 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.095964 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.095974 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.095989 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.095999 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.198059 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.198101 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.198112 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.198127 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.198138 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.300807 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.300865 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.300887 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.300911 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.300927 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.403700 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.403752 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.403768 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.403790 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.403805 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.506235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.506276 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.506287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.506306 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.506317 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.609300 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.609339 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.609350 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.609366 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.609376 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.712291 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.712331 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.712340 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.712355 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.712364 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.815085 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.815128 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.815140 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.815157 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.815169 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.917959 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.917992 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.918003 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.918019 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:12 crc kubenswrapper[4741]: I0929 19:10:12.918031 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:12Z","lastTransitionTime":"2025-09-29T19:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.020693 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.020742 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.020755 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.020773 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.020784 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.085647 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:13 crc kubenswrapper[4741]: E0929 19:10:13.085776 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.086537 4741 scope.go:117] "RemoveContainer" containerID="3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3" Sep 29 19:10:13 crc kubenswrapper[4741]: E0929 19:10:13.086688 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.098940 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.117112 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.123446 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.123483 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.123493 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.123507 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.123516 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.129140 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.141098 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.150803 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.161771 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.175358 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.188132 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.198592 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.215149 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.224258 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.226283 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.226312 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.226324 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.226366 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.226378 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.233545 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.243869 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.255659 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.265122 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.277492 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.285471 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.295696 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.328938 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.329180 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.329254 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.329342 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.329456 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.431967 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.432016 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.432028 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.432043 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.432054 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.534555 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.534589 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.534600 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.534615 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.534625 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.637610 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.637817 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.637876 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.637943 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.638032 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.740754 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.740848 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.740871 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.740954 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.740984 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.843737 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.843804 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.843816 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.843829 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.843838 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.945457 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.945507 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.945515 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.945527 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:13 crc kubenswrapper[4741]: I0929 19:10:13.945535 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:13Z","lastTransitionTime":"2025-09-29T19:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.047335 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.047376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.047403 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.047419 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.047431 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.103656 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:14 crc kubenswrapper[4741]: E0929 19:10:14.103807 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.104140 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:14 crc kubenswrapper[4741]: E0929 19:10:14.104286 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.104562 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:14 crc kubenswrapper[4741]: E0929 19:10:14.104746 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.149918 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.149955 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.149965 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.149980 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.150025 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.252793 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.252841 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.252852 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.252866 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.252876 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.356355 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.356465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.356485 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.356514 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.356537 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.459481 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.459562 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.459589 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.459622 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.459646 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.562376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.562476 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.562495 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.562523 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.562541 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.665807 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.665868 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.665886 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.665912 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.665930 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.769222 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.769272 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.769281 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.769295 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.769304 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.872030 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.872079 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.872092 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.872110 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.872123 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.974995 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.975043 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.975056 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.975072 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:14 crc kubenswrapper[4741]: I0929 19:10:14.975116 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:14Z","lastTransitionTime":"2025-09-29T19:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.078031 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.078292 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.078356 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.078451 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.078514 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.085779 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:15 crc kubenswrapper[4741]: E0929 19:10:15.086893 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.181654 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.181691 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.181703 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.181717 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.181727 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.235148 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.235195 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.235208 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.235225 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.235237 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: E0929 19:10:15.249908 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:15Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.253898 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.253936 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.253948 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.253967 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.253980 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: E0929 19:10:15.266353 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:15Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.269748 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.269906 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.270014 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.270085 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.270141 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: E0929 19:10:15.281605 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:15Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.284879 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.284910 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.284919 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.284931 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.284940 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: E0929 19:10:15.295536 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:15Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.299054 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.299112 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.299126 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.299138 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.299146 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: E0929 19:10:15.309320 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:15Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:15 crc kubenswrapper[4741]: E0929 19:10:15.309666 4741 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.311197 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.311311 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.311423 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.311600 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.311716 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.414214 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.414245 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.414255 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.414267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.414276 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.516970 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.517818 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.517913 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.518055 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.518138 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.620944 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.621201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.621365 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.621491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.621599 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.724677 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.724951 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.725075 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.725161 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.725240 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.828234 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.828554 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.828641 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.828727 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.828800 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.931221 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.931262 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.931281 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.931300 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:15 crc kubenswrapper[4741]: I0929 19:10:15.931311 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:15Z","lastTransitionTime":"2025-09-29T19:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.034009 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.034319 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.034655 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.034783 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.034890 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.084854 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:16 crc kubenswrapper[4741]: E0929 19:10:16.085198 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.084872 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:16 crc kubenswrapper[4741]: E0929 19:10:16.085411 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.084882 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:16 crc kubenswrapper[4741]: E0929 19:10:16.085616 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.137191 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.137249 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.137257 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.137272 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.137282 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.240074 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.240106 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.240113 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.240125 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.240134 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.342235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.342547 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.342611 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.342672 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.342744 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.444814 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.445040 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.445100 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.445158 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.445222 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.547543 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.547573 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.547581 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.547593 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.547602 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.649892 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.650168 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.650260 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.650359 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.650472 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.752438 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.752902 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.753054 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.753180 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.753357 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.855868 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.856126 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.856306 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.856432 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.856521 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.958750 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.958782 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.958793 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.958808 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:16 crc kubenswrapper[4741]: I0929 19:10:16.958818 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:16Z","lastTransitionTime":"2025-09-29T19:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.061434 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.061466 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.061474 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.061487 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.061495 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.085481 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:17 crc kubenswrapper[4741]: E0929 19:10:17.085879 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.164287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.164312 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.164321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.164333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.164340 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.266915 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.266955 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.266964 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.266979 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.266987 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.369165 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.369201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.369210 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.369224 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.369233 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.472535 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.472594 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.472611 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.472639 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.472656 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.575238 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.575298 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.575314 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.575346 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.575385 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.679127 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.679180 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.679196 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.679223 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.679240 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.782576 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.782622 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.782662 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.782681 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.782695 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.885771 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.885817 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.885826 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.885841 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.885851 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.988481 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.988525 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.988548 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.988575 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:17 crc kubenswrapper[4741]: I0929 19:10:17.988590 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:17Z","lastTransitionTime":"2025-09-29T19:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.085491 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.085499 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:18 crc kubenswrapper[4741]: E0929 19:10:18.085641 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:18 crc kubenswrapper[4741]: E0929 19:10:18.085745 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.086030 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:18 crc kubenswrapper[4741]: E0929 19:10:18.086299 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.092493 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.092746 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.092883 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.093019 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.093157 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.197293 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.197340 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.197352 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.197369 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.197383 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.300143 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.300814 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.300920 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.301016 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.301176 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.403808 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.403846 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.403854 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.403868 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.403876 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.506654 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.506698 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.506708 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.506721 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.506730 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.610017 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.610056 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.610067 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.610083 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.610095 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.712546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.712601 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.712623 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.712651 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.712673 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.814929 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.815019 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.815049 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.815085 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.815110 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.918074 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.918138 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.918154 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.918177 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:18 crc kubenswrapper[4741]: I0929 19:10:18.918194 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:18Z","lastTransitionTime":"2025-09-29T19:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.020540 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.020622 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.020638 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.020658 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.020674 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.085226 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:19 crc kubenswrapper[4741]: E0929 19:10:19.085328 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.102693 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.117355 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.122802 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.122841 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.122851 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.122865 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.122875 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.131619 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.144518 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.156218 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.168992 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.184965 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.203567 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.213767 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.224180 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.225544 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.225681 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.225825 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.225950 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.226070 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.236805 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.252216 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.268360 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.287918 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.300154 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.320431 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.328738 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.328769 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.328777 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.328791 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.328800 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.335708 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.372254 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.431497 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.431543 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.431556 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.431575 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.431588 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.534576 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.534625 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.534637 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.534654 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.534668 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.638119 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.638209 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.638237 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.638267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.638287 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.741715 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.741784 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.741796 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.741818 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.741831 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.844178 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.844213 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.844221 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.844235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.844245 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.946871 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.946904 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.946912 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.946926 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:19 crc kubenswrapper[4741]: I0929 19:10:19.946934 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:19Z","lastTransitionTime":"2025-09-29T19:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.049249 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.049291 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.049302 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.049317 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.049327 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.085849 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.085937 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.085954 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:20 crc kubenswrapper[4741]: E0929 19:10:20.086252 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:20 crc kubenswrapper[4741]: E0929 19:10:20.086273 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:20 crc kubenswrapper[4741]: E0929 19:10:20.086642 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.151271 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.151371 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.151499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.151724 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.151747 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.253822 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.253873 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.253888 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.253912 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.253927 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.357766 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.357847 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.357873 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.357907 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.357933 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.461491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.461577 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.461607 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.461666 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.461705 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.564133 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.564192 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.564201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.564215 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.564225 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.667162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.667230 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.667242 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.667261 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.667291 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.770454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.770498 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.770510 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.770528 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.770541 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.872092 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.872130 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.872168 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.872182 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.872193 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.975170 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.975219 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.975233 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.975252 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:20 crc kubenswrapper[4741]: I0929 19:10:20.975267 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:20Z","lastTransitionTime":"2025-09-29T19:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.078111 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.078171 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.078183 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.078202 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.078214 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.085751 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:21 crc kubenswrapper[4741]: E0929 19:10:21.085918 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.180741 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.180808 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.180825 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.180847 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.180865 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.284241 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.284303 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.284315 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.284336 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.284351 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.387358 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.387420 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.387436 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.387454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.387469 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.490012 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.490068 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.490079 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.490105 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.490118 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.592790 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.592928 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.592942 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.592966 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.592985 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.696182 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.696255 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.696276 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.696307 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.696323 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.799635 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.799682 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.799695 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.799711 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.799722 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.902902 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.902963 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.902980 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.903002 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:21 crc kubenswrapper[4741]: I0929 19:10:21.903019 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:21Z","lastTransitionTime":"2025-09-29T19:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.005361 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.005466 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.005486 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.005515 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.005538 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.085679 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.085774 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.085831 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:22 crc kubenswrapper[4741]: E0929 19:10:22.085969 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:22 crc kubenswrapper[4741]: E0929 19:10:22.086318 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:22 crc kubenswrapper[4741]: E0929 19:10:22.086531 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.108557 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.108606 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.108618 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.108637 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.108649 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.211256 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.211300 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.211341 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.211360 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.211372 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.315799 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.315840 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.315854 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.315872 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.315886 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.420080 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.420123 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.420133 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.420153 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.420164 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.523626 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.523695 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.523713 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.523740 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.523758 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.627424 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.627483 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.627499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.627524 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.627541 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.730459 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.730541 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.730568 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.730604 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.730626 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.833297 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.833355 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.833367 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.833384 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.833422 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.936304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.936385 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.936439 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.936465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:22 crc kubenswrapper[4741]: I0929 19:10:22.936486 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:22Z","lastTransitionTime":"2025-09-29T19:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.038639 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.038678 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.038692 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.038710 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.038723 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.085942 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:23 crc kubenswrapper[4741]: E0929 19:10:23.086123 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.141742 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.141802 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.141816 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.141847 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.141863 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.248620 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.248947 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.248992 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.249026 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.249064 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.351968 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.352040 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.352059 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.352087 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.352106 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.455038 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.455144 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.455157 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.455187 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.455200 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.558419 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.558462 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.558474 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.558492 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.558502 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.661608 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.661659 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.661673 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.661696 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.661712 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.765689 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.765742 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.765756 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.765790 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.765810 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.868005 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.868084 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.868103 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.868127 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.868148 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.970793 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.970838 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.970849 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.970864 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:23 crc kubenswrapper[4741]: I0929 19:10:23.970875 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:23Z","lastTransitionTime":"2025-09-29T19:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.030399 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:24 crc kubenswrapper[4741]: E0929 19:10:24.030624 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:10:24 crc kubenswrapper[4741]: E0929 19:10:24.030697 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:10:56.030677986 +0000 UTC m=+97.678467318 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.072516 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.072553 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.072566 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.072581 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.072594 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.085312 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:24 crc kubenswrapper[4741]: E0929 19:10:24.085463 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.085638 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:24 crc kubenswrapper[4741]: E0929 19:10:24.085696 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.085985 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:24 crc kubenswrapper[4741]: E0929 19:10:24.086051 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.174953 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.174987 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.174998 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.175015 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.175026 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.277736 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.277782 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.277795 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.277812 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.277823 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.380145 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.380191 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.380201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.380216 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.380226 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.483346 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.483403 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.483412 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.483425 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.483435 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.586137 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.586642 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.586656 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.586672 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.586682 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.688619 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.688676 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.688692 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.688711 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.688722 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.792048 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.792130 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.792157 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.792194 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.792219 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.895901 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.895954 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.895965 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.895984 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.895997 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.997922 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.997964 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.997975 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.997991 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:24 crc kubenswrapper[4741]: I0929 19:10:24.998079 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:24Z","lastTransitionTime":"2025-09-29T19:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.085736 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:25 crc kubenswrapper[4741]: E0929 19:10:25.085856 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.099959 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.099995 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.100004 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.100017 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.100027 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.201804 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.201841 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.201849 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.201864 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.201872 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.304685 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.304726 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.304737 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.304749 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.304758 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.407277 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.407324 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.407337 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.407355 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.407367 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.510046 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.510100 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.510112 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.510129 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.510145 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.545446 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.545491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.545503 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.545519 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.545532 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: E0929 19:10:25.558818 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:25Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.562225 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.562270 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.562278 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.562291 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.562301 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: E0929 19:10:25.572840 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:25Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.577481 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.577512 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.577521 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.577535 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.577547 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: E0929 19:10:25.589818 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:25Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.592736 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.592776 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.592790 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.592804 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.592813 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: E0929 19:10:25.603957 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:25Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.607222 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.607257 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.607270 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.607287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.607298 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: E0929 19:10:25.621802 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:25Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:25 crc kubenswrapper[4741]: E0929 19:10:25.621963 4741 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.623368 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.623447 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.623466 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.623500 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.623517 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.725745 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.725785 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.725793 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.725807 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.725817 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.827517 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.827548 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.827557 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.827570 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.827578 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.930259 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.930300 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.930320 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.930334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:25 crc kubenswrapper[4741]: I0929 19:10:25.930343 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:25Z","lastTransitionTime":"2025-09-29T19:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.033105 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.033401 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.033503 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.033680 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.033845 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.085118 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:26 crc kubenswrapper[4741]: E0929 19:10:26.085224 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.085572 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.085710 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:26 crc kubenswrapper[4741]: E0929 19:10:26.085873 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:26 crc kubenswrapper[4741]: E0929 19:10:26.086278 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.086535 4741 scope.go:117] "RemoveContainer" containerID="3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.136192 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.136227 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.136238 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.136254 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.136269 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.239236 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.239269 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.239280 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.239295 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.239305 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.341565 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.341599 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.341609 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.341624 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.341636 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.444141 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.444181 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.444192 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.444208 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.444219 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.537768 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/2.log" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.540103 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.540636 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.542240 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/0.log" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.542280 4741 generic.go:334] "Generic (PLEG): container finished" podID="565a367b-9a7a-4b93-b6f8-638b639fef9b" containerID="6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d" exitCode=1 Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.542305 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerDied","Data":"6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.542650 4741 scope.go:117] "RemoveContainer" containerID="6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.545560 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.545574 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.545581 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.545594 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.545602 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.553121 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.571298 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.582371 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.607687 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.617096 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.635422 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.647654 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.647699 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.647710 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.647726 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.647737 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.663065 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.687823 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.702531 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.720588 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.731353 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.743307 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.750375 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.750420 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.750431 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.750444 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.750453 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.754633 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.771749 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.786763 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.798288 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.807023 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.819630 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.831086 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.843159 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.852527 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.852568 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.852577 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.852590 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.852598 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.857091 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.870748 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.884720 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.900539 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.912782 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.933695 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.948500 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.955282 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.955324 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.955332 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.955344 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.955353 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:26Z","lastTransitionTime":"2025-09-29T19:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.963615 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.979081 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:26 crc kubenswrapper[4741]: I0929 19:10:26.993612 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:26Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.008175 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.024688 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.037104 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.050045 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"2025-09-29T19:09:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057\\\\n2025-09-29T19:09:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057 to /host/opt/cni/bin/\\\\n2025-09-29T19:09:41Z [verbose] multus-daemon started\\\\n2025-09-29T19:09:41Z [verbose] Readiness Indicator file check\\\\n2025-09-29T19:10:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.058246 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.058281 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.058291 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.058309 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.058320 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.061305 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.082979 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.085019 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:27 crc kubenswrapper[4741]: E0929 19:10:27.085250 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.161279 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.161778 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.161798 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.161817 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.161829 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.264898 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.264972 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.264983 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.264996 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.265006 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.367344 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.367382 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.367408 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.367421 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.367431 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.470239 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.470557 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.470624 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.470697 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.470758 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.551473 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/0.log" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.551564 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerStarted","Data":"cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.553869 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/3.log" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.554829 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/2.log" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.558011 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" exitCode=1 Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.558047 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.558080 4741 scope.go:117] "RemoveContainer" containerID="3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.558784 4741 scope.go:117] "RemoveContainer" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" Sep 29 19:10:27 crc kubenswrapper[4741]: E0929 19:10:27.558944 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.564943 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.572836 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.572885 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.572897 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.572913 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.572946 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.586036 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.600096 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.619590 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.631234 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.642773 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.654829 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.666807 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.674772 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.675014 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.675087 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.675152 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.675208 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.678897 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.693245 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.704409 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.716485 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"2025-09-29T19:09:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057\\\\n2025-09-29T19:09:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057 to /host/opt/cni/bin/\\\\n2025-09-29T19:09:41Z [verbose] multus-daemon started\\\\n2025-09-29T19:09:41Z [verbose] Readiness Indicator file check\\\\n2025-09-29T19:10:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.732055 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.753446 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.768369 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.778409 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.778448 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.778457 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.778472 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.778482 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.779881 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.791494 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.801956 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.812723 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.826183 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.837555 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.849243 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.862154 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.872202 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.881285 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.881400 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.881476 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.881535 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.881598 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.884941 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.895157 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.912844 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4615416e10b237ae853a070fadae1f5a6ef82258ee298aea511fde9ef5cda3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:01Z\\\",\\\"message\\\":\\\"ions generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193952 6362 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 19:10:01.193982 6362 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reje\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"message\\\":\\\"rieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-controller-manager-operator 74fb4d24-fca9-4903-8f63-e053b6983047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00747f34f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.924351 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.934706 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.945971 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.963004 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.975121 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.983702 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.983881 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.983941 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.984002 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.984069 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:27Z","lastTransitionTime":"2025-09-29T19:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.987094 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:27 crc kubenswrapper[4741]: I0929 19:10:27.996775 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:27Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.008829 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"2025-09-29T19:09:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057\\\\n2025-09-29T19:09:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057 to /host/opt/cni/bin/\\\\n2025-09-29T19:09:41Z [verbose] multus-daemon started\\\\n2025-09-29T19:09:41Z [verbose] Readiness Indicator file check\\\\n2025-09-29T19:10:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.017853 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.085745 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:28 crc kubenswrapper[4741]: E0929 19:10:28.085851 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.086002 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:28 crc kubenswrapper[4741]: E0929 19:10:28.086051 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.086128 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:28 crc kubenswrapper[4741]: E0929 19:10:28.086176 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.087014 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.087046 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.087056 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.087069 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.087078 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.189448 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.189490 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.189499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.189513 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.189524 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.291523 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.291556 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.291568 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.291582 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.291593 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.393926 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.394000 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.394016 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.394038 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.394055 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.496144 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.496184 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.496196 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.496212 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.496222 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.563506 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/3.log" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.567020 4741 scope.go:117] "RemoveContainer" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" Sep 29 19:10:28 crc kubenswrapper[4741]: E0929 19:10:28.567153 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.585407 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.598034 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.598587 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.598678 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.598772 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.598836 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.598898 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.610867 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"2025-09-29T19:09:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057\\\\n2025-09-29T19:09:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057 to /host/opt/cni/bin/\\\\n2025-09-29T19:09:41Z [verbose] multus-daemon started\\\\n2025-09-29T19:09:41Z [verbose] Readiness Indicator file check\\\\n2025-09-29T19:10:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.621566 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.637360 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.650780 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.661494 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.669668 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.677334 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.700957 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.701029 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.701049 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.701072 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.701090 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.709176 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.722869 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.733647 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.749251 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"message\\\":\\\"rieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-controller-manager-operator 74fb4d24-fca9-4903-8f63-e053b6983047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00747f34f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.757975 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.766356 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.777024 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.787551 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.799463 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:28Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.802904 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.802934 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.802944 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.802957 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.802966 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.905186 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.905222 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.905233 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.905251 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:28 crc kubenswrapper[4741]: I0929 19:10:28.905288 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:28Z","lastTransitionTime":"2025-09-29T19:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.007871 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.008156 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.008256 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.008334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.008450 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.085803 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:29 crc kubenswrapper[4741]: E0929 19:10:29.085967 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.100107 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.110279 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.110313 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.110321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.110334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.110342 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.116923 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.131875 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.148846 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"message\\\":\\\"rieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-controller-manager-operator 74fb4d24-fca9-4903-8f63-e053b6983047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00747f34f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.160382 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.172794 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.184185 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.198580 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.210711 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.211953 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.212271 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.212374 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.212497 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.212648 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.223631 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.234559 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.246959 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"2025-09-29T19:09:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057\\\\n2025-09-29T19:09:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057 to /host/opt/cni/bin/\\\\n2025-09-29T19:09:41Z [verbose] multus-daemon started\\\\n2025-09-29T19:09:41Z [verbose] Readiness Indicator file check\\\\n2025-09-29T19:10:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.257707 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.273450 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.287045 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.297071 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.307236 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.314927 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.314962 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.314972 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.314986 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.314996 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.316849 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:29Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.417282 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.417323 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.417334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.417348 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.417358 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.520034 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.520079 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.520097 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.520115 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.520128 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.622579 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.622622 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.622631 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.622646 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.622655 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.724791 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.724815 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.724824 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.724836 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.724844 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.826830 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.826870 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.826879 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.826894 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.826907 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.929789 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.929832 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.929843 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.929858 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:29 crc kubenswrapper[4741]: I0929 19:10:29.929867 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:29Z","lastTransitionTime":"2025-09-29T19:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.032380 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.032457 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.032468 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.032487 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.032500 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.085337 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.085363 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:30 crc kubenswrapper[4741]: E0929 19:10:30.085535 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.085411 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:30 crc kubenswrapper[4741]: E0929 19:10:30.085613 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:30 crc kubenswrapper[4741]: E0929 19:10:30.085704 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.135200 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.135245 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.135254 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.135270 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.135279 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.237789 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.237838 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.237851 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.237870 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.237881 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.340266 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.340313 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.340321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.340334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.340343 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.442341 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.442485 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.442503 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.442610 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.442622 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.545162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.545222 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.545237 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.545257 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.545271 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.647463 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.647505 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.647514 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.647527 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.647536 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.749876 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.749914 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.749927 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.749949 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.749962 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.852523 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.852555 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.852563 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.852595 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.852604 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.955015 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.955075 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.955090 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.955107 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:30 crc kubenswrapper[4741]: I0929 19:10:30.955121 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:30Z","lastTransitionTime":"2025-09-29T19:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.057729 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.057781 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.057792 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.057808 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.057816 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.085545 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:31 crc kubenswrapper[4741]: E0929 19:10:31.085706 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.160479 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.160526 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.160535 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.160547 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.160556 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.262680 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.262709 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.262718 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.262731 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.262740 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.365087 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.365140 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.365158 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.365179 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.365195 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.467001 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.467052 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.467067 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.467084 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.467096 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.569634 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.569674 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.569683 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.569699 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.569709 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.671711 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.671750 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.671758 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.671771 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.671781 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.774077 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.774116 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.774124 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.774138 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.774147 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.876287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.876333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.876343 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.876357 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.876369 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.979027 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.979159 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.979173 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.979186 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:31 crc kubenswrapper[4741]: I0929 19:10:31.979198 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:31Z","lastTransitionTime":"2025-09-29T19:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.081441 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.081480 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.081488 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.081501 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.081510 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.084803 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.084899 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:32 crc kubenswrapper[4741]: E0929 19:10:32.084996 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.085037 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:32 crc kubenswrapper[4741]: E0929 19:10:32.085112 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:32 crc kubenswrapper[4741]: E0929 19:10:32.085205 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.184160 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.184223 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.184240 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.184261 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.184276 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.288231 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.288280 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.288294 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.288311 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.288323 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.391038 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.391121 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.391132 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.391145 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.391156 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.493734 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.493787 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.493803 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.493820 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.493831 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.595643 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.595687 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.595696 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.595709 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.595719 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.698164 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.698246 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.698276 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.698299 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.698315 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.799943 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.799989 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.800000 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.800015 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.800023 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.902427 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.902494 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.902516 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.902538 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:32 crc kubenswrapper[4741]: I0929 19:10:32.902552 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:32Z","lastTransitionTime":"2025-09-29T19:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.005711 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.005792 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.005819 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.005846 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.005863 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.084968 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:33 crc kubenswrapper[4741]: E0929 19:10:33.085092 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.109102 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.109149 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.109165 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.109190 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.109207 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.212081 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.212144 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.212162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.212189 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.212207 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.315796 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.315859 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.315880 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.315904 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.315921 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.419043 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.419149 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.419174 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.419206 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.419226 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.521536 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.521597 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.521607 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.521623 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.521634 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.623833 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.623875 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.623884 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.623901 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.623912 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.726745 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.726808 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.726823 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.726840 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.726853 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.830350 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.830454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.830483 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.830511 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.830528 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.933188 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.933234 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.933245 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.933259 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:33 crc kubenswrapper[4741]: I0929 19:10:33.933268 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:33Z","lastTransitionTime":"2025-09-29T19:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.036117 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.036195 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.036218 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.036247 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.036271 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.084897 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.084989 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:34 crc kubenswrapper[4741]: E0929 19:10:34.085041 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.085097 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:34 crc kubenswrapper[4741]: E0929 19:10:34.085209 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:34 crc kubenswrapper[4741]: E0929 19:10:34.085276 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.139447 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.139513 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.139524 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.139541 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.139551 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.242127 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.242174 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.242187 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.242204 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.242216 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.345895 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.345950 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.345969 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.345994 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.346012 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.448544 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.448596 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.448605 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.448620 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.448630 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.551113 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.551169 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.551185 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.551205 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.551220 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.654361 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.654441 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.654459 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.654487 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.654505 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.758235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.758319 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.758344 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.758376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.758468 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.861884 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.861927 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.861937 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.861950 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.861962 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.964635 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.964706 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.964715 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.964750 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:34 crc kubenswrapper[4741]: I0929 19:10:34.964760 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:34Z","lastTransitionTime":"2025-09-29T19:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.066616 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.066657 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.066668 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.066683 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.066694 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.085596 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:35 crc kubenswrapper[4741]: E0929 19:10:35.085705 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.169626 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.169669 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.169680 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.169695 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.169706 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.272381 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.272445 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.272460 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.272480 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.272492 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.375135 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.375175 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.375187 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.375202 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.375213 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.478205 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.478262 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.478279 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.478304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.478322 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.580447 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.580480 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.580490 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.580501 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.580510 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.683274 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.683370 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.683439 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.683477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.683501 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.787158 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.787244 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.787282 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.787315 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.787337 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.866040 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.866091 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.866103 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.866117 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.866126 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: E0929 19:10:35.884987 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:35Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.889305 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.889357 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.889374 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.889415 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.889432 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: E0929 19:10:35.901879 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:35Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.906788 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.906842 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.906855 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.906874 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.906888 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: E0929 19:10:35.924426 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:35Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.929321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.929366 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.929383 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.929424 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.929440 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: E0929 19:10:35.947378 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:35Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.952533 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.952581 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.952592 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.952609 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.952622 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:35 crc kubenswrapper[4741]: E0929 19:10:35.970447 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:35Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:35 crc kubenswrapper[4741]: E0929 19:10:35.970826 4741 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.972203 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.972260 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.972279 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.972303 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:35 crc kubenswrapper[4741]: I0929 19:10:35.972320 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:35Z","lastTransitionTime":"2025-09-29T19:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.075456 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.075809 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.075977 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.076113 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.076256 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.085048 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.085101 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.085109 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:36 crc kubenswrapper[4741]: E0929 19:10:36.085227 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:36 crc kubenswrapper[4741]: E0929 19:10:36.085421 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:36 crc kubenswrapper[4741]: E0929 19:10:36.085525 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.179232 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.179261 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.179269 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.179281 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.179290 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.282250 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.282316 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.282333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.282358 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.282375 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.385033 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.385086 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.385099 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.385116 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.385126 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.487192 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.487239 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.487248 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.487263 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.487274 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.589363 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.589417 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.589429 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.589464 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.589474 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.691382 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.691428 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.691437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.691449 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.691458 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.793951 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.793981 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.793991 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.794004 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.794013 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.897352 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.897459 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.897499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.897517 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:36 crc kubenswrapper[4741]: I0929 19:10:36.897529 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:36Z","lastTransitionTime":"2025-09-29T19:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.000806 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.000893 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.000919 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.000955 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.000979 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.085566 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:37 crc kubenswrapper[4741]: E0929 19:10:37.085784 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.104316 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.104472 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.104685 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.104704 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.104716 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.208172 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.208225 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.208236 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.208254 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.208266 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.311333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.311433 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.311453 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.311480 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.311496 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.414606 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.414709 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.414737 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.414790 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.414822 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.518089 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.518134 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.518147 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.518163 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.518174 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.620409 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.620482 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.620494 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.620509 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.620521 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.722313 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.722359 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.722375 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.722413 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.722425 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.825012 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.825082 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.825091 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.825107 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.825117 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.928095 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.928167 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.928186 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.928208 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:37 crc kubenswrapper[4741]: I0929 19:10:37.928225 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:37Z","lastTransitionTime":"2025-09-29T19:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.031030 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.031103 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.031137 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.031167 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.031188 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.085113 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.085183 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.085130 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:38 crc kubenswrapper[4741]: E0929 19:10:38.085307 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:38 crc kubenswrapper[4741]: E0929 19:10:38.085444 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:38 crc kubenswrapper[4741]: E0929 19:10:38.085603 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.134178 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.134215 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.134224 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.134238 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.134246 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.236199 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.236235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.236245 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.236259 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.236270 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.338441 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.338472 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.338481 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.338495 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.338506 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.440950 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.440987 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.440998 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.441013 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.441025 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.543068 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.543107 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.543118 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.543131 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.543140 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.644835 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.644891 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.644907 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.644930 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.644948 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.747150 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.747192 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.747205 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.747226 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.747238 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.850072 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.850103 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.850112 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.850126 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.850135 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.953631 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.953664 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.953682 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.953699 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:38 crc kubenswrapper[4741]: I0929 19:10:38.953712 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:38Z","lastTransitionTime":"2025-09-29T19:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.056175 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.056244 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.056268 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.056296 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.056320 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.085781 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:39 crc kubenswrapper[4741]: E0929 19:10:39.085967 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.100226 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.111008 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.156055 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"message\\\":\\\"rieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-controller-manager-operator 74fb4d24-fca9-4903-8f63-e053b6983047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00747f34f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.159210 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.159252 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.159266 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.159285 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.159302 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.176585 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.191725 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.203472 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.214821 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.226865 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.239880 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.252367 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.261346 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.261401 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.261417 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.261432 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.261441 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.265212 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"2025-09-29T19:09:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057\\\\n2025-09-29T19:09:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057 to /host/opt/cni/bin/\\\\n2025-09-29T19:09:41Z [verbose] multus-daemon started\\\\n2025-09-29T19:09:41Z [verbose] Readiness Indicator file check\\\\n2025-09-29T19:10:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.274182 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.291083 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.303882 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.315000 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.325716 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.336600 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.347028 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.364279 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.364753 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.364829 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.364913 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.364990 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.468333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.468374 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.468412 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.468432 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.468443 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.570769 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.570839 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.570856 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.570883 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.570904 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.673477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.673539 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.673559 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.673584 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.673605 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.776275 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.776312 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.776322 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.776337 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.776347 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.878918 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.878970 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.878980 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.878995 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.879005 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.981773 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.981825 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.981837 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.981855 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:39 crc kubenswrapper[4741]: I0929 19:10:39.981870 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:39Z","lastTransitionTime":"2025-09-29T19:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085061 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085119 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085187 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085103 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085250 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085271 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085301 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: E0929 19:10:40.085298 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.085322 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: E0929 19:10:40.085886 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:40 crc kubenswrapper[4741]: E0929 19:10:40.085959 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.086609 4741 scope.go:117] "RemoveContainer" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" Sep 29 19:10:40 crc kubenswrapper[4741]: E0929 19:10:40.086864 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.187929 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.188004 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.188020 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.188043 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.188061 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.290621 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.290670 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.290687 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.290706 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.290722 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.393586 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.393663 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.393682 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.393711 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.393730 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.497028 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.497108 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.497132 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.497167 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.497188 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.600346 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.600465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.600491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.600523 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.600544 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.704546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.704620 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.704641 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.704670 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.704688 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.807465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.807537 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.807565 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.807598 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.807620 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.910716 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.910772 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.910786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.910812 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:40 crc kubenswrapper[4741]: I0929 19:10:40.910827 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:40Z","lastTransitionTime":"2025-09-29T19:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.014162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.014263 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.014289 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.014330 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.014354 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.085008 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.085233 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.117556 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.117603 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.117614 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.117631 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.117644 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.221542 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.221617 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.221641 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.221672 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.221691 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.325104 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.325179 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.325198 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.325227 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.325248 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.429311 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.429380 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.429434 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.429461 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.429480 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.532621 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.532698 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.532717 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.532743 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.532765 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.635825 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.635874 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.635886 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.635903 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.635915 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.739160 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.739208 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.739219 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.739235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.739245 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.833169 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.833345 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.833548 4741 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.833831 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.833804489 +0000 UTC m=+147.481593831 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.833903 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.833943 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.833927023 +0000 UTC m=+147.481716365 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.834076 4741 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.834193 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.834177001 +0000 UTC m=+147.481966343 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.842036 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.842068 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.842077 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.842090 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.842101 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.935159 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.935244 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935466 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935479 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935526 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935550 4741 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935492 4741 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935629 4741 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935655 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.935628801 +0000 UTC m=+147.583418173 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:41 crc kubenswrapper[4741]: E0929 19:10:41.935739 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.935719604 +0000 UTC m=+147.583508956 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.946287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.946435 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.946467 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.946507 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:41 crc kubenswrapper[4741]: I0929 19:10:41.946546 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:41Z","lastTransitionTime":"2025-09-29T19:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.049988 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.050130 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.050158 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.050194 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.050220 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.085705 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:42 crc kubenswrapper[4741]: E0929 19:10:42.085833 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.086062 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:42 crc kubenswrapper[4741]: E0929 19:10:42.086139 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.086252 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:42 crc kubenswrapper[4741]: E0929 19:10:42.086305 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.157287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.157784 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.157799 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.157831 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.157847 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.261268 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.261335 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.261351 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.261375 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.261407 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.364360 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.364412 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.364423 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.364439 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.364449 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.467571 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.467622 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.467635 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.467653 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.467664 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.571759 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.571800 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.571816 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.571831 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.571839 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.674381 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.674477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.674487 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.674502 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.674513 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.777304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.777347 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.777359 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.777376 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.777411 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.880768 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.881066 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.881333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.881624 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.881977 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.984662 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.984699 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.984710 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.984722 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:42 crc kubenswrapper[4741]: I0929 19:10:42.984731 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:42Z","lastTransitionTime":"2025-09-29T19:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.085788 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:43 crc kubenswrapper[4741]: E0929 19:10:43.085937 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.087292 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.087324 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.087336 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.087352 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.087363 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.189703 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.189742 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.189754 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.189769 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.189779 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.292182 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.292234 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.292249 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.292270 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.292286 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.394465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.394500 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.394510 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.394551 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.394560 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.497171 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.497211 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.497219 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.497234 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.497243 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.599882 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.599909 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.599917 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.599929 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.599937 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.702482 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.702527 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.702560 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.702577 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.702588 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.805304 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.805344 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.805353 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.805367 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.805380 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.907838 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.907870 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.907880 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.907895 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:43 crc kubenswrapper[4741]: I0929 19:10:43.907907 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:43Z","lastTransitionTime":"2025-09-29T19:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.010270 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.010332 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.010353 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.010379 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.010438 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.085583 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.085621 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.085681 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:44 crc kubenswrapper[4741]: E0929 19:10:44.085805 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:44 crc kubenswrapper[4741]: E0929 19:10:44.085879 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:44 crc kubenswrapper[4741]: E0929 19:10:44.085956 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.112765 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.112821 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.112855 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.112886 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.112911 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.215095 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.215147 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.215163 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.215185 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.215203 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.317760 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.317800 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.317808 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.317822 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.317833 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.420282 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.420308 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.420316 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.420329 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.420337 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.522944 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.522987 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.523002 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.523022 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.523041 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.625612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.625818 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.625835 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.625860 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.625875 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.729856 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.729937 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.729951 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.729974 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.729991 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.832911 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.832981 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.832999 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.833026 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.833047 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.936333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.936442 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.936462 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.936488 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:44 crc kubenswrapper[4741]: I0929 19:10:44.936509 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:44Z","lastTransitionTime":"2025-09-29T19:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.039437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.039485 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.039497 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.039517 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.039530 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.085622 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:45 crc kubenswrapper[4741]: E0929 19:10:45.085833 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.142612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.142676 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.142690 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.142712 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.142727 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.246546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.246612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.246626 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.246645 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.246662 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.349436 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.349483 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.349499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.349517 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.349531 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.453335 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.453429 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.453444 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.453477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.453496 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.557211 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.557295 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.557321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.557361 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.557386 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.660257 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.660309 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.660322 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.660368 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.660383 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.763299 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.763337 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.763348 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.763363 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.763375 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.866096 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.866146 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.866158 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.866182 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.866198 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.969037 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.969093 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.969107 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.969124 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:45 crc kubenswrapper[4741]: I0929 19:10:45.969139 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:45Z","lastTransitionTime":"2025-09-29T19:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.071333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.071381 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.071430 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.071453 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.071470 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.085252 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.085512 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.085786 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.085924 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.085979 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.086121 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.151449 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.151491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.151500 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.151517 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.151527 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.170444 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.176062 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.176134 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.176145 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.176164 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.176177 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.196004 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.202314 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.202440 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.202472 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.202512 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.202545 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.220639 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.225674 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.225744 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.225775 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.225809 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.225835 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.248259 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.254260 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.254331 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.254355 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.254385 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.254435 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.270720 4741 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8fb5d538-81ea-402d-ba81-35b3f9382fe2\\\",\\\"systemUUID\\\":\\\"43f60388-9284-4af8-8654-468a0103453b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:46Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:46 crc kubenswrapper[4741]: E0929 19:10:46.270877 4741 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.273342 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.273400 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.273421 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.273446 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.273462 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.376507 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.376555 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.376591 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.376613 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.376624 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.480454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.480511 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.480525 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.480549 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.480567 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.584725 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.584786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.584796 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.584812 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.584822 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.687237 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.687294 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.687307 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.687325 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.687336 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.790862 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.790913 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.790923 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.790940 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.790952 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.894649 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.894745 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.894769 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.894804 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.894830 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.999129 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.999186 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.999197 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.999214 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:46 crc kubenswrapper[4741]: I0929 19:10:46.999225 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:46Z","lastTransitionTime":"2025-09-29T19:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.086295 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:47 crc kubenswrapper[4741]: E0929 19:10:47.086565 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.103939 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.104710 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.104771 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.104789 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.104815 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.104835 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.208134 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.208213 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.208226 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.208247 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.208266 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.311485 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.311546 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.311589 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.311621 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.311640 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.415299 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.415432 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.415462 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.415497 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.415568 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.518832 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.518917 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.518937 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.518970 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.518991 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.622704 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.622746 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.622755 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.622769 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.622779 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.725139 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.725201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.725212 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.725227 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.725238 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.827786 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.827853 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.827872 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.827898 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.827918 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.931228 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.931372 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.931422 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.931454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:47 crc kubenswrapper[4741]: I0929 19:10:47.931474 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:47Z","lastTransitionTime":"2025-09-29T19:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.033875 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.033942 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.033961 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.033993 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.034019 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.085288 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.085290 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.085323 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:48 crc kubenswrapper[4741]: E0929 19:10:48.085589 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:48 crc kubenswrapper[4741]: E0929 19:10:48.085716 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:48 crc kubenswrapper[4741]: E0929 19:10:48.085857 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.138306 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.138364 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.138378 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.138429 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.138440 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.242056 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.242151 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.242178 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.242211 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.242235 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.345238 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.345328 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.345351 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.345380 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.345439 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.447845 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.447891 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.447902 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.447916 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.447926 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.550597 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.550672 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.550689 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.550714 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.550731 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.653161 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.653211 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.653223 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.653240 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.653252 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.756416 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.756494 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.756510 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.756538 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.756557 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.860020 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.860079 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.860092 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.860113 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.860128 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.963283 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.963350 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.963368 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.963420 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:48 crc kubenswrapper[4741]: I0929 19:10:48.963437 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:48Z","lastTransitionTime":"2025-09-29T19:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.066353 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.066428 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.066440 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.066453 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.066462 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.085312 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:49 crc kubenswrapper[4741]: E0929 19:10:49.085515 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.124578 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d23f1763-eee3-4ba3-8fcb-6533a9b1e67b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df1ce2602a28d80c061a2af0fa2d601e26cf5c4b101eb2ab404f2b65dd376e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5b66ca3dc6df756e812dfcfb103b825c7c4f642860228a1c9011e56ba44006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09ceae42e893099e43b26ceca5e069eb886c65f971a10ce1df6b48e1d925a1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://074950bf812df689c720f3eda51b7a6ae957452c58515a33faf62d938efc35a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8fddbf6f6c361edef4661184fbda87df8fd83834d36b718f61841c435530df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2cb1a1370b7227ff4d7eb26a76fa4c1497bde6db3b6a8d3b3100a03328bfd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58a8f96cdbf4775a8a8abcf759dc5c46705e934985721421fcaa1e2734aa76b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ffadd875926799a3f1bbfc6b1ce2825a827c63d6fcf451eb3f85e2d5c9ead93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.143312 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"636d93d8-0c53-436c-9f41-089cbd6106c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df3ddff9b3ee2de1e396f6ad90567e0ebf8f46aa2ce7f5dda8f620adeb986e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49c62cd1709ba1eb278cf78d4cb6a97612342b55178e69fc7b2569c108cd165\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e1c929ef1698111704d41701584bcef423157fdafa40e861116409c67fa454\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a28e7d12f33440efc482f28e14c79870e1c43c9503f8d2af810f380885af0c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa231498b7c56724893479f52c941f751733555c431eab274f41b7af867158\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T19:09:37Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 19:09:22.440457 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 19:09:22.441539 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3209846921/tls.crt::/tmp/serving-cert-3209846921/tls.key\\\\\\\"\\\\nI0929 19:09:37.941065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 19:09:37.945297 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 19:09:37.945318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 19:09:37.945336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 19:09:37.945342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 19:09:37.951914 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 19:09:37.951944 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 19:09:37.951952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 19:09:37.951955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 19:09:37.951958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 19:09:37.951960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 19:09:37.952175 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 19:09:37.959725 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31ac0867813dfcd4baa1b0bc199fb2d9d280ec74acce8651b05274159b2f7074\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff12bc0e283fa4ba632c929edff5055b73b655de289f303150e4925e0875ab05\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.159569 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64d0f51caf6ef308113cdcd3df3997c3469140871028a7c25a94d5826a075139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae01afe5dabb74e6bea3c34ae5dc04b533f226cac5faee21e12bdd0231981d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.169712 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.169753 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.169767 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.169788 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.169799 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.173242 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nz2l9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27362580-9e01-4eb6-be43-1372e7131232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2cf09e3614d44787a24da6c7853a9eceb3813bb5b7bbddc19caad5bd2322ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbz86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nz2l9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.196499 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ql4kz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"565a367b-9a7a-4b93-b6f8-638b639fef9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:26Z\\\",\\\"message\\\":\\\"2025-09-29T19:09:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057\\\\n2025-09-29T19:09:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d14a6053-88c9-4064-84e6-ea42ae0d5057 to /host/opt/cni/bin/\\\\n2025-09-29T19:09:41Z [verbose] multus-daemon started\\\\n2025-09-29T19:09:41Z [verbose] Readiness Indicator file check\\\\n2025-09-29T19:10:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlt5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ql4kz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.209799 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388588a6-1c5b-4478-8895-40997b0e00c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fd75bda43426dbd2c61b58acd89342b9e284a3781acdc011b913b469b75d289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e7378701e804f06792ea60b9a14b1a7a794e8fdf06a66e2160565fd61e8912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wkrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlr7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.226198 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b1f4011-9296-420b-8a8a-48ea15dbd99c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef753504cb83d924ff41cdcd47083961d35ebf43fe4be7293ffcccc24c09592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34f84a30634f0e985236321731e959a593236609ce6ca33acdb94a2aab138018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34f84a30634f0e985236321731e959a593236609ce6ca33acdb94a2aab138018\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.240848 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40036c2f-4a32-4b42-9fc9-dcc60acca237\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a27193087d1010dc28a2f4fa99d7c43cb146a80ed68bdfd10c9c73448c5b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2993ac4bbf04969a871a45c9715ae911455b78f9d6e3768aaf13a9feecf319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5736b2430e52ef0c772d2d2ad55f66e4c29b000e1caf290402a3c0ce96e0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4eb7f261eae3f2dee80e090fb23f3e74d1c236bb9788fc57c9495ac2224173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.259064 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.271635 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.271668 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.271680 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.271698 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.271709 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.272497 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.283226 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ebda18-417a-46ae-93ed-035b301fd203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2618a82fa13425aac65b2c25ef6268725183019c8b465803274c3df5bcfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://671c59478ef9e9f004186476002635e7072cb86a2946ca4720fcbaccb44b4faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a93a73a9d801d2ad3d1a185fcf684542485e8e82602d0fa197f5a69cb4b0b05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceef0de642b0f7abb381464fed4455d81e8cde20285fbf009648517e93ba22a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.295322 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf8291492ed0dd1f3148a6717d41b2e484fd41e01f7fc35d5dad43f6e5b9b5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.306065 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e4b008175221e02ed65c2b0ceadf843ea91134550a97b01d8b5d2bcc7fd6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.318720 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13be2fb8-f24c-4a58-827a-3d0a5bf8dc72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d65d98ad1e93382c73c1609c3056134b93fe271973b48956047108619ce34b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cd98446b595e487866d064ec3128a8a558c62737adcd5b2d8b505c3c7cfdfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54d8a53d3e8c9a50abeb7c6da070e16fb9496d4e5eee93815a2eb8ded9cc74a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0d17e0ed1a664a5ca374bc973a657df0d026b4fce5880670de0e01c17cf0c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1d363be5134b94fd012779201fd55328940b6864c1359666eb1f0ba796014c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7aa57ba91ec9f07a2e41eb0c682ed1c535faf9d5e90a41c40e3723b280edef7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0c932db13be6332dca7e983cd143147e79bd1a06b1214c34facd1ea8059295a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8sz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8tbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.328452 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2f20aa2-390a-494e-aed0-0dfd0e031d3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f890ae0637ba803db5ba2d318a508ccdf25ed519d96f79745faf1bdc4523b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsl74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rtxqk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.348303 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a31e12c-13a1-4052-b679-7115e1a3574e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T19:10:27Z\\\",\\\"message\\\":\\\"rieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-controller-manager-operator 74fb4d24-fca9-4903-8f63-e053b6983047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00747f34f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T19:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T19:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T19:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhpnc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mrdpd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.360375 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fwxc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95b724d8-e45d-4f98-86be-85df7d4a57da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ac9c6168a9328a27f8dcbcb4bce4841b393a05cf86ffb076e48e770c05731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T19:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztjnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fwxc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.370756 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7krvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58436159-f44b-45ad-98f9-a60bc389bc7b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmf6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T19:09:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7krvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.373864 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.373910 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.373925 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.373940 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.373952 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.382002 4741 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T19:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T19:10:49Z is after 2025-08-24T17:21:41Z" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.476809 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.476870 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.476889 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.476908 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.476920 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.579171 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.579263 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.579285 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.579319 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.579361 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.682685 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.682731 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.682741 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.682759 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.682770 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.786231 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.786279 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.786289 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.786301 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.786310 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.888994 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.889040 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.889052 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.889067 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.889078 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.991486 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.991567 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.991581 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.991597 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:49 crc kubenswrapper[4741]: I0929 19:10:49.991608 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:49Z","lastTransitionTime":"2025-09-29T19:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.085610 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.085672 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:50 crc kubenswrapper[4741]: E0929 19:10:50.085724 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:50 crc kubenswrapper[4741]: E0929 19:10:50.085839 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.085696 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:50 crc kubenswrapper[4741]: E0929 19:10:50.085945 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.093384 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.093450 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.093465 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.093477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.093488 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.195981 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.196019 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.196028 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.196041 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.196053 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.298855 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.298893 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.298901 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.298916 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.298924 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.401224 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.401281 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.401298 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.401316 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.401328 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.504232 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.504303 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.505829 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.505873 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.505892 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.609702 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.610061 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.610321 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.610435 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.610528 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.714004 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.714087 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.714102 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.714121 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.714136 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.816105 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.816142 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.816153 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.816170 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.816182 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.919277 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.919325 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.919414 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.919436 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:50 crc kubenswrapper[4741]: I0929 19:10:50.919449 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:50Z","lastTransitionTime":"2025-09-29T19:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.022433 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.022499 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.022526 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.022557 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.022579 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.084871 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:51 crc kubenswrapper[4741]: E0929 19:10:51.085045 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.086103 4741 scope.go:117] "RemoveContainer" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" Sep 29 19:10:51 crc kubenswrapper[4741]: E0929 19:10:51.086326 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.125651 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.126029 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.126103 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.126178 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.126322 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.228139 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.228176 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.228185 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.228198 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.228207 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.330559 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.330591 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.330600 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.330612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.330622 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.433194 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.433248 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.433260 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.433278 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.433290 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.540267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.540310 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.540329 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.540350 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.540362 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.643317 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.643431 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.643458 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.643491 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.643514 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.746066 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.746129 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.746143 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.746162 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.746176 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.848796 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.848847 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.848858 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.848874 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.848886 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.951287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.951324 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.951333 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.951346 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:51 crc kubenswrapper[4741]: I0929 19:10:51.951355 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:51Z","lastTransitionTime":"2025-09-29T19:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.054185 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.054235 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.054250 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.054267 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.054278 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.085636 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.085747 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.085655 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:52 crc kubenswrapper[4741]: E0929 19:10:52.085868 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:52 crc kubenswrapper[4741]: E0929 19:10:52.085938 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:52 crc kubenswrapper[4741]: E0929 19:10:52.086022 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.157377 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.157454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.157463 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.157477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.157488 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.262212 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.262338 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.262348 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.262363 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.262372 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.364598 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.364636 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.364644 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.364661 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.364671 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.466884 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.466998 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.467030 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.467061 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.467086 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.571294 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.571377 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.571454 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.571497 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.571519 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.674064 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.674108 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.674119 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.674136 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.674149 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.776769 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.776807 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.776822 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.776838 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.776848 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.879703 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.879766 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.879788 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.879818 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.879840 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.981856 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.981927 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.981952 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.981979 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:52 crc kubenswrapper[4741]: I0929 19:10:52.981999 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:52Z","lastTransitionTime":"2025-09-29T19:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.084140 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.084173 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.084180 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.084193 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.084202 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.085479 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:53 crc kubenswrapper[4741]: E0929 19:10:53.085580 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.186211 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.186249 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.186256 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.186270 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.186286 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.289238 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.289296 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.289312 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.289334 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.289350 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.392010 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.392246 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.392264 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.392298 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.392312 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.495293 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.495364 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.495411 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.495437 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.495454 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.597572 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.597626 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.597640 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.597658 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.597670 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.700233 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.700314 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.700339 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.700372 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.700449 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.803142 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.803179 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.803188 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.803201 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.803212 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.905576 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.905616 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.905627 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.905643 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:53 crc kubenswrapper[4741]: I0929 19:10:53.905654 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:53Z","lastTransitionTime":"2025-09-29T19:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.008521 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.008550 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.008560 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.008576 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.008587 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.084983 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.085074 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:54 crc kubenswrapper[4741]: E0929 19:10:54.085113 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.084983 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:54 crc kubenswrapper[4741]: E0929 19:10:54.085193 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:54 crc kubenswrapper[4741]: E0929 19:10:54.085441 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.110938 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.110989 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.111012 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.111033 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.111049 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.213204 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.213249 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.213258 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.213274 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.213287 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.316308 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.316361 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.316372 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.316404 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.316416 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.418958 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.419002 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.419015 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.419037 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.419054 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.521528 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.521581 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.521592 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.521612 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.521622 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.623848 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.623900 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.623918 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.623938 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.623953 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.726859 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.726902 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.726913 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.726928 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.726939 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.829179 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.829225 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.829236 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.829253 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.829268 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.931541 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.931609 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.931629 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.931651 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:54 crc kubenswrapper[4741]: I0929 19:10:54.931672 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:54Z","lastTransitionTime":"2025-09-29T19:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.034434 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.034477 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.034489 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.034508 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.034520 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.085429 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:55 crc kubenswrapper[4741]: E0929 19:10:55.085580 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.137298 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.137348 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.137364 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.137384 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.137417 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.239589 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.239624 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.239636 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.239658 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.239674 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.342315 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.342373 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.342407 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.342429 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.342441 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.445466 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.445530 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.445547 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.445569 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.445584 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.548875 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.548919 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.548928 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.548945 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.548954 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.650883 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.650914 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.650923 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.650936 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.650952 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.752845 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.752884 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.752893 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.752905 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.752916 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.855146 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.855185 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.855202 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.855217 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.855227 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.957569 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.957631 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.957647 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.957668 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:55 crc kubenswrapper[4741]: I0929 19:10:55.957682 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:55Z","lastTransitionTime":"2025-09-29T19:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.060287 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.060330 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.060342 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.060356 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.060367 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:56Z","lastTransitionTime":"2025-09-29T19:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.085591 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.085700 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.085720 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:56 crc kubenswrapper[4741]: E0929 19:10:56.085839 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:56 crc kubenswrapper[4741]: E0929 19:10:56.085973 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:56 crc kubenswrapper[4741]: E0929 19:10:56.086073 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.089937 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:56 crc kubenswrapper[4741]: E0929 19:10:56.090063 4741 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:10:56 crc kubenswrapper[4741]: E0929 19:10:56.090107 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs podName:58436159-f44b-45ad-98f9-a60bc389bc7b nodeName:}" failed. No retries permitted until 2025-09-29 19:12:00.090093902 +0000 UTC m=+161.737883224 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs") pod "network-metrics-daemon-7krvl" (UID: "58436159-f44b-45ad-98f9-a60bc389bc7b") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.162819 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.162857 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.162869 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.162884 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.162895 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:56Z","lastTransitionTime":"2025-09-29T19:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.265567 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.265613 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.265622 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.265637 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.265652 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:56Z","lastTransitionTime":"2025-09-29T19:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.368099 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.368139 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.368152 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.368167 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.368178 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:56Z","lastTransitionTime":"2025-09-29T19:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.470749 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.470791 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.470800 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.470817 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.470826 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:56Z","lastTransitionTime":"2025-09-29T19:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.551896 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.551936 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.551952 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.551971 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.551983 4741 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T19:10:56Z","lastTransitionTime":"2025-09-29T19:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.595463 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj"] Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.595816 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.596992 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.598218 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.598343 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.600011 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.627347 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.627330739 podStartE2EDuration="1m15.627330739s" podCreationTimestamp="2025-09-29 19:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.627238766 +0000 UTC m=+98.275028128" watchObservedRunningTime="2025-09-29 19:10:56.627330739 +0000 UTC m=+98.275120071" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.670356 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-q8tbb" podStartSLOduration=78.670340007 podStartE2EDuration="1m18.670340007s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.670069049 +0000 UTC m=+98.317858401" watchObservedRunningTime="2025-09-29 19:10:56.670340007 +0000 UTC m=+98.318129339" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.696012 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32877316-0616-4718-9ed4-476472fb48dc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.696067 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/32877316-0616-4718-9ed4-476472fb48dc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.696747 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/32877316-0616-4718-9ed4-476472fb48dc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.696783 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32877316-0616-4718-9ed4-476472fb48dc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.696817 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32877316-0616-4718-9ed4-476472fb48dc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.716347 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podStartSLOduration=78.716327808 podStartE2EDuration="1m18.716327808s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.685576184 +0000 UTC m=+98.333365546" watchObservedRunningTime="2025-09-29 19:10:56.716327808 +0000 UTC m=+98.364117150" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.726244 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-fwxc2" podStartSLOduration=77.726222429 podStartE2EDuration="1m17.726222429s" podCreationTimestamp="2025-09-29 19:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.725573768 +0000 UTC m=+98.373363110" watchObservedRunningTime="2025-09-29 19:10:56.726222429 +0000 UTC m=+98.374011771" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.774858 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=75.774839582 podStartE2EDuration="1m15.774839582s" podCreationTimestamp="2025-09-29 19:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.760926876 +0000 UTC m=+98.408716218" watchObservedRunningTime="2025-09-29 19:10:56.774839582 +0000 UTC m=+98.422628914" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.775377 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.775368949 podStartE2EDuration="1m18.775368949s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.774724648 +0000 UTC m=+98.422513980" watchObservedRunningTime="2025-09-29 19:10:56.775368949 +0000 UTC m=+98.423158281" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.797756 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/32877316-0616-4718-9ed4-476472fb48dc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.797797 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/32877316-0616-4718-9ed4-476472fb48dc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.797824 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32877316-0616-4718-9ed4-476472fb48dc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.797851 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32877316-0616-4718-9ed4-476472fb48dc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.797909 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32877316-0616-4718-9ed4-476472fb48dc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.798054 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/32877316-0616-4718-9ed4-476472fb48dc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.798121 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/32877316-0616-4718-9ed4-476472fb48dc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.798530 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-nz2l9" podStartSLOduration=79.798520434 podStartE2EDuration="1m19.798520434s" podCreationTimestamp="2025-09-29 19:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.797854063 +0000 UTC m=+98.445643405" watchObservedRunningTime="2025-09-29 19:10:56.798520434 +0000 UTC m=+98.446309756" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.798872 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32877316-0616-4718-9ed4-476472fb48dc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.806128 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32877316-0616-4718-9ed4-476472fb48dc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.811425 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-ql4kz" podStartSLOduration=78.811406338 podStartE2EDuration="1m18.811406338s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.810359596 +0000 UTC m=+98.458148928" watchObservedRunningTime="2025-09-29 19:10:56.811406338 +0000 UTC m=+98.459195670" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.819059 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32877316-0616-4718-9ed4-476472fb48dc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xmcnj\" (UID: \"32877316-0616-4718-9ed4-476472fb48dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.822331 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlr7m" podStartSLOduration=78.82231236 podStartE2EDuration="1m18.82231236s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.822045572 +0000 UTC m=+98.469834904" watchObservedRunningTime="2025-09-29 19:10:56.82231236 +0000 UTC m=+98.470101692" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.854702 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.854680435 podStartE2EDuration="47.854680435s" podCreationTimestamp="2025-09-29 19:10:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.854572461 +0000 UTC m=+98.502361793" watchObservedRunningTime="2025-09-29 19:10:56.854680435 +0000 UTC m=+98.502469777" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.855826 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=9.855818751 podStartE2EDuration="9.855818751s" podCreationTimestamp="2025-09-29 19:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:56.843333509 +0000 UTC m=+98.491122841" watchObservedRunningTime="2025-09-29 19:10:56.855818751 +0000 UTC m=+98.503608093" Sep 29 19:10:56 crc kubenswrapper[4741]: I0929 19:10:56.911553 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" Sep 29 19:10:56 crc kubenswrapper[4741]: W0929 19:10:56.927685 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32877316_0616_4718_9ed4_476472fb48dc.slice/crio-9b0b2cc809e5d19ad21874f93636ed93d911841bcde2f6e8c9f8b00a51d4abe9 WatchSource:0}: Error finding container 9b0b2cc809e5d19ad21874f93636ed93d911841bcde2f6e8c9f8b00a51d4abe9: Status 404 returned error can't find the container with id 9b0b2cc809e5d19ad21874f93636ed93d911841bcde2f6e8c9f8b00a51d4abe9 Sep 29 19:10:57 crc kubenswrapper[4741]: I0929 19:10:57.084978 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:57 crc kubenswrapper[4741]: E0929 19:10:57.085105 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:10:57 crc kubenswrapper[4741]: I0929 19:10:57.660614 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" event={"ID":"32877316-0616-4718-9ed4-476472fb48dc","Type":"ContainerStarted","Data":"26ea042b4213606a9678af60befda9fc0e4d627fa2449c9558add960d33a6373"} Sep 29 19:10:57 crc kubenswrapper[4741]: I0929 19:10:57.660665 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" event={"ID":"32877316-0616-4718-9ed4-476472fb48dc","Type":"ContainerStarted","Data":"9b0b2cc809e5d19ad21874f93636ed93d911841bcde2f6e8c9f8b00a51d4abe9"} Sep 29 19:10:57 crc kubenswrapper[4741]: I0929 19:10:57.678514 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xmcnj" podStartSLOduration=79.678475233 podStartE2EDuration="1m19.678475233s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:10:57.677913525 +0000 UTC m=+99.325702857" watchObservedRunningTime="2025-09-29 19:10:57.678475233 +0000 UTC m=+99.326264565" Sep 29 19:10:58 crc kubenswrapper[4741]: I0929 19:10:58.085737 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:10:58 crc kubenswrapper[4741]: I0929 19:10:58.085786 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:10:58 crc kubenswrapper[4741]: I0929 19:10:58.085764 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:10:58 crc kubenswrapper[4741]: E0929 19:10:58.085898 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:10:58 crc kubenswrapper[4741]: E0929 19:10:58.086010 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:10:58 crc kubenswrapper[4741]: E0929 19:10:58.086098 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:10:59 crc kubenswrapper[4741]: I0929 19:10:59.085573 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:10:59 crc kubenswrapper[4741]: E0929 19:10:59.087037 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:00 crc kubenswrapper[4741]: I0929 19:11:00.085294 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:00 crc kubenswrapper[4741]: I0929 19:11:00.085344 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:00 crc kubenswrapper[4741]: E0929 19:11:00.085412 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:00 crc kubenswrapper[4741]: I0929 19:11:00.085478 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:00 crc kubenswrapper[4741]: E0929 19:11:00.085661 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:00 crc kubenswrapper[4741]: E0929 19:11:00.085687 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:01 crc kubenswrapper[4741]: I0929 19:11:01.085498 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:01 crc kubenswrapper[4741]: E0929 19:11:01.085648 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:02 crc kubenswrapper[4741]: I0929 19:11:02.085929 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:02 crc kubenswrapper[4741]: I0929 19:11:02.086009 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:02 crc kubenswrapper[4741]: E0929 19:11:02.086108 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:02 crc kubenswrapper[4741]: I0929 19:11:02.086196 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:02 crc kubenswrapper[4741]: E0929 19:11:02.086345 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:02 crc kubenswrapper[4741]: E0929 19:11:02.086885 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:03 crc kubenswrapper[4741]: I0929 19:11:03.085765 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:03 crc kubenswrapper[4741]: E0929 19:11:03.086002 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:03 crc kubenswrapper[4741]: I0929 19:11:03.087801 4741 scope.go:117] "RemoveContainer" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" Sep 29 19:11:03 crc kubenswrapper[4741]: E0929 19:11:03.088330 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mrdpd_openshift-ovn-kubernetes(7a31e12c-13a1-4052-b679-7115e1a3574e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" Sep 29 19:11:04 crc kubenswrapper[4741]: I0929 19:11:04.084958 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:04 crc kubenswrapper[4741]: I0929 19:11:04.085056 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:04 crc kubenswrapper[4741]: I0929 19:11:04.085184 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:04 crc kubenswrapper[4741]: E0929 19:11:04.085176 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:04 crc kubenswrapper[4741]: E0929 19:11:04.085321 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:04 crc kubenswrapper[4741]: E0929 19:11:04.085366 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:05 crc kubenswrapper[4741]: I0929 19:11:05.085525 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:05 crc kubenswrapper[4741]: E0929 19:11:05.085827 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:06 crc kubenswrapper[4741]: I0929 19:11:06.086050 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:06 crc kubenswrapper[4741]: I0929 19:11:06.086114 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:06 crc kubenswrapper[4741]: I0929 19:11:06.086114 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:06 crc kubenswrapper[4741]: E0929 19:11:06.086325 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:06 crc kubenswrapper[4741]: E0929 19:11:06.086572 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:06 crc kubenswrapper[4741]: E0929 19:11:06.086813 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:07 crc kubenswrapper[4741]: I0929 19:11:07.085859 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:07 crc kubenswrapper[4741]: E0929 19:11:07.086049 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:08 crc kubenswrapper[4741]: I0929 19:11:08.084945 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:08 crc kubenswrapper[4741]: E0929 19:11:08.085121 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:08 crc kubenswrapper[4741]: I0929 19:11:08.085273 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:08 crc kubenswrapper[4741]: I0929 19:11:08.085450 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:08 crc kubenswrapper[4741]: E0929 19:11:08.085485 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:08 crc kubenswrapper[4741]: E0929 19:11:08.085623 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:09 crc kubenswrapper[4741]: I0929 19:11:09.085810 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:09 crc kubenswrapper[4741]: E0929 19:11:09.089349 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:10 crc kubenswrapper[4741]: I0929 19:11:10.085520 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:10 crc kubenswrapper[4741]: I0929 19:11:10.085526 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:10 crc kubenswrapper[4741]: E0929 19:11:10.086114 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:10 crc kubenswrapper[4741]: I0929 19:11:10.085595 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:10 crc kubenswrapper[4741]: E0929 19:11:10.086235 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:10 crc kubenswrapper[4741]: E0929 19:11:10.086503 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:11 crc kubenswrapper[4741]: I0929 19:11:11.085278 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:11 crc kubenswrapper[4741]: E0929 19:11:11.085417 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.085134 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.085194 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.085253 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:12 crc kubenswrapper[4741]: E0929 19:11:12.085407 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:12 crc kubenswrapper[4741]: E0929 19:11:12.085494 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:12 crc kubenswrapper[4741]: E0929 19:11:12.085585 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.710436 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/1.log" Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.711126 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/0.log" Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.711177 4741 generic.go:334] "Generic (PLEG): container finished" podID="565a367b-9a7a-4b93-b6f8-638b639fef9b" containerID="cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1" exitCode=1 Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.711206 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerDied","Data":"cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1"} Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.711241 4741 scope.go:117] "RemoveContainer" containerID="6037630326c3a9c2c33fdcf898010d7a4ee476d0a048c365487356cca043be1d" Sep 29 19:11:12 crc kubenswrapper[4741]: I0929 19:11:12.712308 4741 scope.go:117] "RemoveContainer" containerID="cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1" Sep 29 19:11:12 crc kubenswrapper[4741]: E0929 19:11:12.712634 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-ql4kz_openshift-multus(565a367b-9a7a-4b93-b6f8-638b639fef9b)\"" pod="openshift-multus/multus-ql4kz" podUID="565a367b-9a7a-4b93-b6f8-638b639fef9b" Sep 29 19:11:13 crc kubenswrapper[4741]: I0929 19:11:13.085554 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:13 crc kubenswrapper[4741]: E0929 19:11:13.085852 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:13 crc kubenswrapper[4741]: I0929 19:11:13.717512 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/1.log" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.085865 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.085868 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.085928 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:14 crc kubenswrapper[4741]: E0929 19:11:14.086109 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:14 crc kubenswrapper[4741]: E0929 19:11:14.086180 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:14 crc kubenswrapper[4741]: E0929 19:11:14.086338 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.087542 4741 scope.go:117] "RemoveContainer" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.723466 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/3.log" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.726819 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerStarted","Data":"3637cdd09ff11428d033a392b90d98862680635742030c3645664937ffd7b6b2"} Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.727616 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.754488 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podStartSLOduration=96.754462436 podStartE2EDuration="1m36.754462436s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:14.753579718 +0000 UTC m=+116.401369050" watchObservedRunningTime="2025-09-29 19:11:14.754462436 +0000 UTC m=+116.402251768" Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.993987 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7krvl"] Sep 29 19:11:14 crc kubenswrapper[4741]: I0929 19:11:14.994076 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:14 crc kubenswrapper[4741]: E0929 19:11:14.994149 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:15 crc kubenswrapper[4741]: I0929 19:11:15.085988 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:15 crc kubenswrapper[4741]: E0929 19:11:15.086178 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:16 crc kubenswrapper[4741]: I0929 19:11:16.084906 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:16 crc kubenswrapper[4741]: I0929 19:11:16.085083 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:16 crc kubenswrapper[4741]: E0929 19:11:16.085319 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:16 crc kubenswrapper[4741]: I0929 19:11:16.085400 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:16 crc kubenswrapper[4741]: E0929 19:11:16.085612 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:16 crc kubenswrapper[4741]: E0929 19:11:16.086141 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:17 crc kubenswrapper[4741]: I0929 19:11:17.085602 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:17 crc kubenswrapper[4741]: E0929 19:11:17.085765 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:18 crc kubenswrapper[4741]: I0929 19:11:18.085296 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:18 crc kubenswrapper[4741]: I0929 19:11:18.085383 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:18 crc kubenswrapper[4741]: E0929 19:11:18.085467 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:18 crc kubenswrapper[4741]: E0929 19:11:18.085532 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:18 crc kubenswrapper[4741]: I0929 19:11:18.085312 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:18 crc kubenswrapper[4741]: E0929 19:11:18.085618 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:19 crc kubenswrapper[4741]: I0929 19:11:19.085349 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:19 crc kubenswrapper[4741]: E0929 19:11:19.087672 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:19 crc kubenswrapper[4741]: E0929 19:11:19.111090 4741 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Sep 29 19:11:19 crc kubenswrapper[4741]: E0929 19:11:19.251347 4741 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 19:11:20 crc kubenswrapper[4741]: I0929 19:11:20.085428 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:20 crc kubenswrapper[4741]: I0929 19:11:20.085478 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:20 crc kubenswrapper[4741]: I0929 19:11:20.085566 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:20 crc kubenswrapper[4741]: E0929 19:11:20.085956 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:20 crc kubenswrapper[4741]: E0929 19:11:20.086094 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:20 crc kubenswrapper[4741]: E0929 19:11:20.086220 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:21 crc kubenswrapper[4741]: I0929 19:11:21.085960 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:21 crc kubenswrapper[4741]: E0929 19:11:21.086121 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:22 crc kubenswrapper[4741]: I0929 19:11:22.085584 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:22 crc kubenswrapper[4741]: I0929 19:11:22.085621 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:22 crc kubenswrapper[4741]: I0929 19:11:22.085705 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:22 crc kubenswrapper[4741]: E0929 19:11:22.085721 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:22 crc kubenswrapper[4741]: E0929 19:11:22.085849 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:22 crc kubenswrapper[4741]: E0929 19:11:22.086017 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:23 crc kubenswrapper[4741]: I0929 19:11:23.085353 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:23 crc kubenswrapper[4741]: E0929 19:11:23.086383 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:24 crc kubenswrapper[4741]: I0929 19:11:24.087344 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:24 crc kubenswrapper[4741]: I0929 19:11:24.087364 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:24 crc kubenswrapper[4741]: E0929 19:11:24.087548 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:24 crc kubenswrapper[4741]: I0929 19:11:24.087400 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:24 crc kubenswrapper[4741]: E0929 19:11:24.087633 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:24 crc kubenswrapper[4741]: E0929 19:11:24.087827 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:24 crc kubenswrapper[4741]: E0929 19:11:24.253045 4741 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 19:11:25 crc kubenswrapper[4741]: I0929 19:11:25.086005 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:25 crc kubenswrapper[4741]: E0929 19:11:25.086165 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:26 crc kubenswrapper[4741]: I0929 19:11:26.084914 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:26 crc kubenswrapper[4741]: I0929 19:11:26.084951 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:26 crc kubenswrapper[4741]: I0929 19:11:26.085127 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:26 crc kubenswrapper[4741]: E0929 19:11:26.085320 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:26 crc kubenswrapper[4741]: I0929 19:11:26.085345 4741 scope.go:117] "RemoveContainer" containerID="cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1" Sep 29 19:11:26 crc kubenswrapper[4741]: E0929 19:11:26.085547 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:26 crc kubenswrapper[4741]: E0929 19:11:26.085752 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:26 crc kubenswrapper[4741]: I0929 19:11:26.797276 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/1.log" Sep 29 19:11:26 crc kubenswrapper[4741]: I0929 19:11:26.797349 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerStarted","Data":"472e7bb64bf7085f0ba716a2fc6cc1eb722f462abeeec7bbfddd1fc9cfcc3941"} Sep 29 19:11:27 crc kubenswrapper[4741]: I0929 19:11:27.085696 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:27 crc kubenswrapper[4741]: E0929 19:11:27.086508 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:28 crc kubenswrapper[4741]: I0929 19:11:28.085723 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:28 crc kubenswrapper[4741]: I0929 19:11:28.085801 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:28 crc kubenswrapper[4741]: I0929 19:11:28.085815 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:28 crc kubenswrapper[4741]: E0929 19:11:28.085970 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 19:11:28 crc kubenswrapper[4741]: E0929 19:11:28.086074 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7krvl" podUID="58436159-f44b-45ad-98f9-a60bc389bc7b" Sep 29 19:11:28 crc kubenswrapper[4741]: E0929 19:11:28.086169 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 19:11:29 crc kubenswrapper[4741]: I0929 19:11:29.084950 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:29 crc kubenswrapper[4741]: E0929 19:11:29.086293 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.084873 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.084892 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.084892 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.088328 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.088462 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.088682 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.088851 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.088971 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.089133 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Sep 29 19:11:30 crc kubenswrapper[4741]: I0929 19:11:30.354408 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:11:31 crc kubenswrapper[4741]: I0929 19:11:31.085062 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.377840 4741 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.417084 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.417445 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.419662 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.419942 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.431868 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-jd7m4"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.432219 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.449138 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.449517 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.449886 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.452620 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dgp6j"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.452933 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.453320 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.454407 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.454530 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.455038 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.455953 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.456068 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.458740 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.459109 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.460958 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.461332 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.461553 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.461800 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.462022 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.462205 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.463617 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.464968 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.465090 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.467438 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.471494 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.471540 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.471623 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.471681 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.471724 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.471855 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pxhts"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.471884 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.472402 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.472759 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jtclz"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.473247 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.473554 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nrp5d"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.473717 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.473883 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.474030 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.474068 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.474118 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.476413 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.476613 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xjn62"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.478722 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.480160 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-d4ltl"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.480508 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27cn9"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.480672 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.480788 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.480938 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.481056 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.481434 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.481825 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.481882 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.481828 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.482341 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.482615 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.482635 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.482841 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.482978 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.483175 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.483421 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.483565 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.483725 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.489791 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.490249 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8xdlm"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.490461 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4r5pc"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.490728 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.491341 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.491575 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8xdlm" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.492797 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.493059 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.493190 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.493340 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.493472 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.493576 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.493683 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.493915 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.494082 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.494270 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.494410 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.496112 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.496453 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.496664 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497243 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497294 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h8d49"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497437 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497453 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497590 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497856 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497859 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.498194 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.498212 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.498240 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497910 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497990 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.497258 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.499293 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.500687 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.501110 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.510507 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.511829 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.512630 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.512919 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.513053 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.513251 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.513426 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.513458 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.513677 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.514013 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.514739 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.514743 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.515269 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.515699 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.541105 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.541685 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.541695 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.541896 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.553221 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.553605 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.553714 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.553828 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.553914 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554013 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554063 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.553924 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554134 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554153 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554162 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.553832 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554383 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554517 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.554633 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.556274 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.556299 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.556758 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.556932 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.557094 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.557349 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.557491 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.558146 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.558612 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.558953 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.559172 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.560006 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.560187 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.560822 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.560846 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.562780 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.564056 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.564637 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.565245 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.566063 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.567587 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hktj8"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.568201 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.570688 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571127 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571786 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ff7353-843d-41bf-86eb-96a315f31bc3-config\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571821 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b277\" (UniqueName: \"kubernetes.io/projected/8bd18131-53d9-4d13-a85b-41366f624859-kube-api-access-8b277\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571844 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-client-ca\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571864 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/013d0fb9-848e-4156-a67f-498daa95fc6f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571888 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgxmt\" (UniqueName: \"kubernetes.io/projected/013d0fb9-848e-4156-a67f-498daa95fc6f-kube-api-access-qgxmt\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571930 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71ee9d0-0827-4cac-aa2f-a6835259d37b-serving-cert\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571945 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kws77\" (UniqueName: \"kubernetes.io/projected/c71ee9d0-0827-4cac-aa2f-a6835259d37b-kube-api-access-kws77\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571961 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-config\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571978 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-etcd-ca\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.571995 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1ff7353-843d-41bf-86eb-96a315f31bc3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572014 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qjgl\" (UniqueName: \"kubernetes.io/projected/a1ff7353-843d-41bf-86eb-96a315f31bc3-kube-api-access-9qjgl\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572034 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8gzx\" (UniqueName: \"kubernetes.io/projected/28c53633-9e71-4d02-859e-0cdca912a924-kube-api-access-h8gzx\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572195 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bd18131-53d9-4d13-a85b-41366f624859-auth-proxy-config\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572331 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8bd18131-53d9-4d13-a85b-41366f624859-machine-approver-tls\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572414 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bd18131-53d9-4d13-a85b-41366f624859-config\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572482 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1ff7353-843d-41bf-86eb-96a315f31bc3-images\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572596 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-config\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572618 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/28c53633-9e71-4d02-859e-0cdca912a924-etcd-client\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572641 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/013d0fb9-848e-4156-a67f-498daa95fc6f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572666 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-etcd-service-ca\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.572734 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28c53633-9e71-4d02-859e-0cdca912a924-serving-cert\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.579649 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.585759 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.586031 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ww5bb"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.586801 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.589110 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.589888 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.592308 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.592501 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.593008 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.593432 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.593676 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.593382 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fjfzz"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.621186 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.623556 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.632785 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.633615 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.637706 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-jd7m4"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.637741 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.638276 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q5rwg"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.638571 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.638852 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzlbd"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.638931 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.639234 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.640005 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.640054 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.640004 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.641459 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dgp6j"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.641570 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.642124 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-f99d7"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.643952 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.645016 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-z95st"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.649299 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.649539 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.650937 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pxhts"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.666790 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.671372 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-d4ltl"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.675450 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8xdlm"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.675525 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.675536 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4r5pc"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.679762 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.679810 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.679822 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hktj8"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680726 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ff7353-843d-41bf-86eb-96a315f31bc3-config\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680764 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b277\" (UniqueName: \"kubernetes.io/projected/8bd18131-53d9-4d13-a85b-41366f624859-kube-api-access-8b277\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680785 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-client-ca\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680805 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/013d0fb9-848e-4156-a67f-498daa95fc6f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680820 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgxmt\" (UniqueName: \"kubernetes.io/projected/013d0fb9-848e-4156-a67f-498daa95fc6f-kube-api-access-qgxmt\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680847 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71ee9d0-0827-4cac-aa2f-a6835259d37b-serving-cert\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680862 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kws77\" (UniqueName: \"kubernetes.io/projected/c71ee9d0-0827-4cac-aa2f-a6835259d37b-kube-api-access-kws77\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680878 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-config\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680892 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-etcd-ca\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680908 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1ff7353-843d-41bf-86eb-96a315f31bc3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680924 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qjgl\" (UniqueName: \"kubernetes.io/projected/a1ff7353-843d-41bf-86eb-96a315f31bc3-kube-api-access-9qjgl\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680946 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8gzx\" (UniqueName: \"kubernetes.io/projected/28c53633-9e71-4d02-859e-0cdca912a924-kube-api-access-h8gzx\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680968 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bd18131-53d9-4d13-a85b-41366f624859-auth-proxy-config\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.680984 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8bd18131-53d9-4d13-a85b-41366f624859-machine-approver-tls\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.681001 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bd18131-53d9-4d13-a85b-41366f624859-config\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.681015 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1ff7353-843d-41bf-86eb-96a315f31bc3-images\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.681039 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-config\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.681055 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/28c53633-9e71-4d02-859e-0cdca912a924-etcd-client\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.681069 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/013d0fb9-848e-4156-a67f-498daa95fc6f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.681087 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-etcd-service-ca\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.681115 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28c53633-9e71-4d02-859e-0cdca912a924-serving-cert\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.682854 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.683495 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-client-ca\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.683831 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/013d0fb9-848e-4156-a67f-498daa95fc6f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.684150 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-etcd-service-ca\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.684585 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ff7353-843d-41bf-86eb-96a315f31bc3-config\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.684940 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bd18131-53d9-4d13-a85b-41366f624859-config\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.684991 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-config\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.685430 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bd18131-53d9-4d13-a85b-41366f624859-auth-proxy-config\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.685841 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-config\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.685875 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/28c53633-9e71-4d02-859e-0cdca912a924-etcd-ca\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.685920 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1ff7353-843d-41bf-86eb-96a315f31bc3-images\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.687115 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1ff7353-843d-41bf-86eb-96a315f31bc3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.687143 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71ee9d0-0827-4cac-aa2f-a6835259d37b-serving-cert\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.687367 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/28c53633-9e71-4d02-859e-0cdca912a924-etcd-client\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.687636 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/013d0fb9-848e-4156-a67f-498daa95fc6f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.688263 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xjn62"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.689735 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28c53633-9e71-4d02-859e-0cdca912a924-serving-cert\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.690025 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8bd18131-53d9-4d13-a85b-41366f624859-machine-approver-tls\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.694989 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.696000 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jtclz"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.696945 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q5rwg"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.698584 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nrp5d"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.700623 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.700814 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.701798 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.702900 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.705764 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.707364 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.708506 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h8d49"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.710012 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.711030 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.712146 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.713441 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.714078 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.715664 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzlbd"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.716664 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-wn992"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.717161 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.717684 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.719175 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.719890 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27cn9"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.720913 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-f99d7"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.721993 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fjfzz"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.723174 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-z95st"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.724335 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.725957 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.727351 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.728373 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-c6qb4"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.729747 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.729757 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-c6qb4"] Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.740102 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.759121 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.779691 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.805343 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.820029 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.839751 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.859615 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.880162 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.919434 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.940068 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.959459 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.979807 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Sep 29 19:11:37 crc kubenswrapper[4741]: I0929 19:11:37.999584 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.020602 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.039335 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.059638 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.079208 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.099173 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.119674 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.140186 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.159867 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.180278 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.199573 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.220489 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.239852 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.260252 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.279871 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.300835 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.319899 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.360032 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.379368 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387032 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mndq4\" (UniqueName: \"kubernetes.io/projected/61b6a0fd-fa98-4401-8a8f-96f0415124e7-kube-api-access-mndq4\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387118 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2985ed-ad44-4d45-9b9b-ad1438579635-trusted-ca\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387148 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-audit\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387172 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3168a724-8ec2-46b6-b151-38e733138a95-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387194 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-config\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387250 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-trusted-ca-bundle\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387303 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61b6a0fd-fa98-4401-8a8f-96f0415124e7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387364 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-certificates\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387384 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2985ed-ad44-4d45-9b9b-ad1438579635-serving-cert\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387448 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5q8m\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-kube-api-access-p5q8m\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387464 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9336afb7-78e3-4ad1-8a48-8401f628d060-serving-cert\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387526 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clcjk\" (UniqueName: \"kubernetes.io/projected/bd769734-8dc9-445d-b482-5e8ef9357a0d-kube-api-access-clcjk\") pod \"downloads-7954f5f757-8xdlm\" (UID: \"bd769734-8dc9-445d-b482-5e8ef9357a0d\") " pod="openshift-console/downloads-7954f5f757-8xdlm" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387562 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-oauth-serving-cert\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387604 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wccqb\" (UniqueName: \"kubernetes.io/projected/5e2985ed-ad44-4d45-9b9b-ad1438579635-kube-api-access-wccqb\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387634 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-ca-trust-extracted\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387728 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387766 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387831 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.387917 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-trusted-ca\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388107 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-etcd-serving-ca\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388130 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/68bd7878-4ed3-42e6-859b-9763ebdf8435-metrics-tls\") pod \"dns-operator-744455d44c-jtclz\" (UID: \"68bd7878-4ed3-42e6-859b-9763ebdf8435\") " pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388145 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3168a724-8ec2-46b6-b151-38e733138a95-config\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388160 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e2985ed-ad44-4d45-9b9b-ad1438579635-config\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388180 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-oauth-config\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388209 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9336afb7-78e3-4ad1-8a48-8401f628d060-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388226 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-config\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388242 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/61b6a0fd-fa98-4401-8a8f-96f0415124e7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388262 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0efc688-e298-42c1-94c6-c56049a7ec9b-config\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388281 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-audit-policies\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388296 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgsk\" (UniqueName: \"kubernetes.io/projected/acdf5b58-ab2f-4526-893c-b9685a63e8a5-kube-api-access-plgsk\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388314 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-serving-cert\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388329 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c26ln\" (UniqueName: \"kubernetes.io/projected/8036b4af-58a3-4e55-9643-4832021a71c5-kube-api-access-c26ln\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388347 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-config\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.388373 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:38.888360768 +0000 UTC m=+140.536150100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388423 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-bound-sa-token\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388443 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-serving-cert\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388458 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0efc688-e298-42c1-94c6-c56049a7ec9b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388473 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmbc\" (UniqueName: \"kubernetes.io/projected/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-kube-api-access-vdmbc\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388487 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd5l9\" (UniqueName: \"kubernetes.io/projected/3168a724-8ec2-46b6-b151-38e733138a95-kube-api-access-rd5l9\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388501 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-etcd-client\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388515 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61b6a0fd-fa98-4401-8a8f-96f0415124e7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388532 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-etcd-client\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388545 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-audit-dir\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388560 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5jc9\" (UniqueName: \"kubernetes.io/projected/9336afb7-78e3-4ad1-8a48-8401f628d060-kube-api-access-m5jc9\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388575 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-service-ca-bundle\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388605 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-config\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388619 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr9sb\" (UniqueName: \"kubernetes.io/projected/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-kube-api-access-qr9sb\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388638 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-encryption-config\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388665 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvskl\" (UniqueName: \"kubernetes.io/projected/68bd7878-4ed3-42e6-859b-9763ebdf8435-kube-api-access-qvskl\") pod \"dns-operator-744455d44c-jtclz\" (UID: \"68bd7878-4ed3-42e6-859b-9763ebdf8435\") " pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388684 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388697 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-encryption-config\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388713 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8036b4af-58a3-4e55-9643-4832021a71c5-node-pullsecrets\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388740 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-serving-cert\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388754 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-image-import-ca\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388766 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-serving-cert\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388786 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-service-ca\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388803 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8036b4af-58a3-4e55-9643-4832021a71c5-audit-dir\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388827 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmr9q\" (UniqueName: \"kubernetes.io/projected/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-kube-api-access-wmr9q\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388887 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-client-ca\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388905 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388928 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-tls\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388942 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acdf5b58-ab2f-4526-893c-b9685a63e8a5-serving-cert\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388957 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388972 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-installation-pull-secrets\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.388986 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0efc688-e298-42c1-94c6-c56049a7ec9b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.399079 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.420320 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.440585 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.459455 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.480019 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.489466 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.489599 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:38.989572178 +0000 UTC m=+140.637361520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.489821 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmrm9\" (UniqueName: \"kubernetes.io/projected/2c152df8-1048-49bb-b981-4853855487d3-kube-api-access-lmrm9\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.489908 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b87bc33f-e3e7-420d-98de-e648fc610aec-config\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.489999 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-serving-cert\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490082 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-image-import-ca\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490176 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-plugins-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490319 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjgtj\" (UniqueName: \"kubernetes.io/projected/7b4710bf-98cc-4ac1-9356-da1249819e88-kube-api-access-qjgtj\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490437 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c77h\" (UniqueName: \"kubernetes.io/projected/c374c787-dd0b-472d-914f-e51d1b63ef89-kube-api-access-2c77h\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490539 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490645 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490732 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c152df8-1048-49bb-b981-4853855487d3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490838 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/88c5801d-74cf-48c9-8443-8dcd8a743ff3-metrics-tls\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.490943 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b87bc33f-e3e7-420d-98de-e648fc610aec-serving-cert\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.491059 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.491126 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-image-import-ca\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.491235 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.491345 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/98c86725-7ab9-4054-9006-2e666f5b9576-images\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.491473 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-policies\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.491584 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-tls\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.491706 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acdf5b58-ab2f-4526-893c-b9685a63e8a5-serving-cert\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492250 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-metrics-certs\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492343 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-installation-pull-secrets\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492493 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492383 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492577 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50ffdc3-0fc9-447f-af03-14b6ce63568d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492724 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2985ed-ad44-4d45-9b9b-ad1438579635-trusted-ca\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492758 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-config\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492788 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3168a724-8ec2-46b6-b151-38e733138a95-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492821 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/965617c1-858f-4aa3-8c89-185f45434afd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492853 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4235e5f-4b25-4e97-b13d-57fcde51d267-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hktj8\" (UID: \"a4235e5f-4b25-4e97-b13d-57fcde51d267\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492883 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-mountpoint-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492913 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931ec446-cf7a-4040-ae73-05a6ba6c327f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492941 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9wpd\" (UniqueName: \"kubernetes.io/projected/807a87dc-a461-46d7-b116-ca81d06d88a1-kube-api-access-s9wpd\") pod \"package-server-manager-789f6589d5-866mj\" (UID: \"807a87dc-a461-46d7-b116-ca81d06d88a1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492971 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d50ffdc3-0fc9-447f-af03-14b6ce63568d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.492999 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-certificates\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493029 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/965617c1-858f-4aa3-8c89-185f45434afd-metrics-tls\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493056 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5q8m\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-kube-api-access-p5q8m\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493081 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7bwb\" (UniqueName: \"kubernetes.io/projected/98c86725-7ab9-4054-9006-2e666f5b9576-kube-api-access-k7bwb\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493114 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493141 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clcjk\" (UniqueName: \"kubernetes.io/projected/bd769734-8dc9-445d-b482-5e8ef9357a0d-kube-api-access-clcjk\") pod \"downloads-7954f5f757-8xdlm\" (UID: \"bd769734-8dc9-445d-b482-5e8ef9357a0d\") " pod="openshift-console/downloads-7954f5f757-8xdlm" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493168 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-oauth-serving-cert\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493194 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wccqb\" (UniqueName: \"kubernetes.io/projected/5e2985ed-ad44-4d45-9b9b-ad1438579635-kube-api-access-wccqb\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493219 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-certs\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493252 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6394941-7bc5-45db-9404-6fa52d0fb10f-config-volume\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493271 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-config\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493289 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493334 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-etcd-serving-ca\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493375 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4a3602e4-822b-4451-90ea-44eaeed085c5-tmpfs\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493412 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9g5d\" (UniqueName: \"kubernetes.io/projected/6e3c7165-5c43-4b86-9428-0b0c2846ef7b-kube-api-access-k9g5d\") pod \"ingress-canary-z95st\" (UID: \"6e3c7165-5c43-4b86-9428-0b0c2846ef7b\") " pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493433 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cq8x\" (UniqueName: \"kubernetes.io/projected/965617c1-858f-4aa3-8c89-185f45434afd-kube-api-access-6cq8x\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493454 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bfpv\" (UniqueName: \"kubernetes.io/projected/7030a932-f4cf-4ef3-9db5-9e3a7da7a547-kube-api-access-9bfpv\") pod \"migrator-59844c95c7-25hcn\" (UID: \"7030a932-f4cf-4ef3-9db5-9e3a7da7a547\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493474 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-config\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493491 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-audit-policies\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493508 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493529 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-bound-sa-token\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493544 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-serving-cert\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493559 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c26ln\" (UniqueName: \"kubernetes.io/projected/8036b4af-58a3-4e55-9643-4832021a71c5-kube-api-access-c26ln\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493580 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-serving-cert\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493600 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd5l9\" (UniqueName: \"kubernetes.io/projected/3168a724-8ec2-46b6-b151-38e733138a95-kube-api-access-rd5l9\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493618 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-stats-auth\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493636 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12453f09-2b49-47af-9f73-8c18c93a1812-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5zmm\" (UID: \"12453f09-2b49-47af-9f73-8c18c93a1812\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493654 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-etcd-client\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493672 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-etcd-client\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493694 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61b6a0fd-fa98-4401-8a8f-96f0415124e7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.493711 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:38.993692101 +0000 UTC m=+140.641481443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493744 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6m6\" (UniqueName: \"kubernetes.io/projected/a4235e5f-4b25-4e97-b13d-57fcde51d267-kube-api-access-7c6m6\") pod \"multus-admission-controller-857f4d67dd-hktj8\" (UID: \"a4235e5f-4b25-4e97-b13d-57fcde51d267\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493776 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88c5801d-74cf-48c9-8443-8dcd8a743ff3-config-volume\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493799 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/98c86725-7ab9-4054-9006-2e666f5b9576-proxy-tls\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493830 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5jc9\" (UniqueName: \"kubernetes.io/projected/9336afb7-78e3-4ad1-8a48-8401f628d060-kube-api-access-m5jc9\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493862 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-service-ca-bundle\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493872 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2985ed-ad44-4d45-9b9b-ad1438579635-trusted-ca\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493920 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-encryption-config\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493941 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr9sb\" (UniqueName: \"kubernetes.io/projected/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-kube-api-access-qr9sb\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.493986 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/199ce9b1-1552-4cd4-a84f-67af5391ab73-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-267cw\" (UID: \"199ce9b1-1552-4cd4-a84f-67af5391ab73\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494006 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-865jv\" (UniqueName: \"kubernetes.io/projected/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-kube-api-access-865jv\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494023 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8036b4af-58a3-4e55-9643-4832021a71c5-node-pullsecrets\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494055 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-serving-cert\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494071 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494088 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c152df8-1048-49bb-b981-4853855487d3-srv-cert\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494107 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8036b4af-58a3-4e55-9643-4832021a71c5-audit-dir\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494125 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-service-ca\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494141 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-default-certificate\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494160 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz8jj\" (UniqueName: \"kubernetes.io/projected/12453f09-2b49-47af-9f73-8c18c93a1812-kube-api-access-qz8jj\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5zmm\" (UID: \"12453f09-2b49-47af-9f73-8c18c93a1812\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494174 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6e3c7165-5c43-4b86-9428-0b0c2846ef7b-cert\") pod \"ingress-canary-z95st\" (UID: \"6e3c7165-5c43-4b86-9428-0b0c2846ef7b\") " pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494201 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmr9q\" (UniqueName: \"kubernetes.io/projected/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-kube-api-access-wmr9q\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494217 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-client-ca\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494234 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjgbd\" (UniqueName: \"kubernetes.io/projected/199ce9b1-1552-4cd4-a84f-67af5391ab73-kube-api-access-sjgbd\") pod \"cluster-samples-operator-665b6dd947-267cw\" (UID: \"199ce9b1-1552-4cd4-a84f-67af5391ab73\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494254 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/931ec446-cf7a-4040-ae73-05a6ba6c327f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494268 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a3602e4-822b-4451-90ea-44eaeed085c5-webhook-cert\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494287 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494302 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494319 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-registration-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494319 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-etcd-serving-ca\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494348 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494380 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a3602e4-822b-4451-90ea-44eaeed085c5-apiservice-cert\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494428 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494462 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0efc688-e298-42c1-94c6-c56049a7ec9b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494488 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-node-bootstrap-token\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494515 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-socket-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494518 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-tls\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494540 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494565 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22vt5\" (UniqueName: \"kubernetes.io/projected/4a3602e4-822b-4451-90ea-44eaeed085c5-kube-api-access-22vt5\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494609 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mndq4\" (UniqueName: \"kubernetes.io/projected/61b6a0fd-fa98-4401-8a8f-96f0415124e7-kube-api-access-mndq4\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494631 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-csi-data-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494670 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/807a87dc-a461-46d7-b116-ca81d06d88a1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-866mj\" (UID: \"807a87dc-a461-46d7-b116-ca81d06d88a1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494692 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/98c86725-7ab9-4054-9006-2e666f5b9576-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494724 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgmsj\" (UniqueName: \"kubernetes.io/projected/b6394941-7bc5-45db-9404-6fa52d0fb10f-kube-api-access-vgmsj\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494751 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn5wc\" (UniqueName: \"kubernetes.io/projected/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-kube-api-access-nn5wc\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494778 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-audit\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494800 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7s8r\" (UniqueName: \"kubernetes.io/projected/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-kube-api-access-p7s8r\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494822 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljmbs\" (UniqueName: \"kubernetes.io/projected/88c5801d-74cf-48c9-8443-8dcd8a743ff3-kube-api-access-ljmbs\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494849 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-trusted-ca-bundle\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494869 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61b6a0fd-fa98-4401-8a8f-96f0415124e7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494891 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494914 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2985ed-ad44-4d45-9b9b-ad1438579635-serving-cert\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494936 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50ffdc3-0fc9-447f-af03-14b6ce63568d-config\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494960 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9336afb7-78e3-4ad1-8a48-8401f628d060-serving-cert\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.494986 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c374c787-dd0b-472d-914f-e51d1b63ef89-signing-cabundle\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495020 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-ca-trust-extracted\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495042 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495067 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495095 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kksm6\" (UniqueName: \"kubernetes.io/projected/d3f43d0d-188d-4a54-9800-dd480764d67d-kube-api-access-kksm6\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495118 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495122 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-trusted-ca\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495145 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/68bd7878-4ed3-42e6-859b-9763ebdf8435-metrics-tls\") pod \"dns-operator-744455d44c-jtclz\" (UID: \"68bd7878-4ed3-42e6-859b-9763ebdf8435\") " pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495215 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c374c787-dd0b-472d-914f-e51d1b63ef89-signing-key\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495249 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-oauth-config\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495273 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3168a724-8ec2-46b6-b151-38e733138a95-config\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495295 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e2985ed-ad44-4d45-9b9b-ad1438579635-config\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495318 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495351 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9336afb7-78e3-4ad1-8a48-8401f628d060-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495368 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495372 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/61b6a0fd-fa98-4401-8a8f-96f0415124e7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495456 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0efc688-e298-42c1-94c6-c56049a7ec9b-config\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495488 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-dir\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495527 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgsk\" (UniqueName: \"kubernetes.io/projected/acdf5b58-ab2f-4526-893c-b9685a63e8a5-kube-api-access-plgsk\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495555 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/931ec446-cf7a-4040-ae73-05a6ba6c327f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495592 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-config\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495615 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0efc688-e298-42c1-94c6-c56049a7ec9b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495644 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmbc\" (UniqueName: \"kubernetes.io/projected/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-kube-api-access-vdmbc\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495671 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495696 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6lgr\" (UniqueName: \"kubernetes.io/projected/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-kube-api-access-j6lgr\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495718 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjggg\" (UniqueName: \"kubernetes.io/projected/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-kube-api-access-hjggg\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495743 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-audit-dir\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495769 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/965617c1-858f-4aa3-8c89-185f45434afd-trusted-ca\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495795 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6394941-7bc5-45db-9404-6fa52d0fb10f-secret-volume\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495917 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f43d0d-188d-4a54-9800-dd480764d67d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495944 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-srv-cert\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.495985 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-config\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496012 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496037 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-proxy-tls\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496062 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvskl\" (UniqueName: \"kubernetes.io/projected/68bd7878-4ed3-42e6-859b-9763ebdf8435-kube-api-access-qvskl\") pod \"dns-operator-744455d44c-jtclz\" (UID: \"68bd7878-4ed3-42e6-859b-9763ebdf8435\") " pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496104 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496130 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-encryption-config\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496154 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b4710bf-98cc-4ac1-9356-da1249819e88-service-ca-bundle\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496181 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22j5p\" (UniqueName: \"kubernetes.io/projected/9c4347d9-37c2-4c1e-806b-1c2927c1638b-kube-api-access-22j5p\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496220 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3f43d0d-188d-4a54-9800-dd480764d67d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.496247 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxbtt\" (UniqueName: \"kubernetes.io/projected/b87bc33f-e3e7-420d-98de-e648fc610aec-kube-api-access-sxbtt\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.497605 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-encryption-config\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.497632 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-certificates\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.497717 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8036b4af-58a3-4e55-9643-4832021a71c5-node-pullsecrets\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.498153 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/61b6a0fd-fa98-4401-8a8f-96f0415124e7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.498808 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8036b4af-58a3-4e55-9643-4832021a71c5-audit\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.499031 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-serving-cert\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.499289 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-config\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.499491 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3168a724-8ec2-46b6-b151-38e733138a95-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.499893 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0efc688-e298-42c1-94c6-c56049a7ec9b-config\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.499941 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8036b4af-58a3-4e55-9643-4832021a71c5-audit-dir\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.500050 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-client-ca\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.500087 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-serving-cert\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.500374 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acdf5b58-ab2f-4526-893c-b9685a63e8a5-serving-cert\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.500569 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-service-ca\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.500777 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9336afb7-78e3-4ad1-8a48-8401f628d060-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.500841 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3168a724-8ec2-46b6-b151-38e733138a95-config\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.502102 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-ca-trust-extracted\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.501554 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.501593 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/68bd7878-4ed3-42e6-859b-9763ebdf8435-metrics-tls\") pod \"dns-operator-744455d44c-jtclz\" (UID: \"68bd7878-4ed3-42e6-859b-9763ebdf8435\") " pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.501708 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e2985ed-ad44-4d45-9b9b-ad1438579635-config\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.501707 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-config\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.501405 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acdf5b58-ab2f-4526-893c-b9685a63e8a5-service-ca-bundle\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.502368 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-audit-dir\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.502517 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-trusted-ca\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.502660 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.503069 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-audit-policies\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.504372 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-config\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.504536 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-oauth-serving-cert\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.504705 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-trusted-ca-bundle\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.505516 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61b6a0fd-fa98-4401-8a8f-96f0415124e7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.505889 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-encryption-config\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.506006 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-etcd-client\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.506206 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.506728 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8036b4af-58a3-4e55-9643-4832021a71c5-etcd-client\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.507264 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-serving-cert\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.507543 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-installation-pull-secrets\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.507635 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-serving-cert\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.508189 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2985ed-ad44-4d45-9b9b-ad1438579635-serving-cert\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.508980 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-oauth-config\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.509153 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9336afb7-78e3-4ad1-8a48-8401f628d060-serving-cert\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.509707 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0efc688-e298-42c1-94c6-c56049a7ec9b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.519541 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.539682 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.559790 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.580451 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.597866 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.598135 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.098106693 +0000 UTC m=+140.745896025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598245 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-stats-auth\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598296 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12453f09-2b49-47af-9f73-8c18c93a1812-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5zmm\" (UID: \"12453f09-2b49-47af-9f73-8c18c93a1812\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598351 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6m6\" (UniqueName: \"kubernetes.io/projected/a4235e5f-4b25-4e97-b13d-57fcde51d267-kube-api-access-7c6m6\") pod \"multus-admission-controller-857f4d67dd-hktj8\" (UID: \"a4235e5f-4b25-4e97-b13d-57fcde51d267\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598462 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88c5801d-74cf-48c9-8443-8dcd8a743ff3-config-volume\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598539 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/98c86725-7ab9-4054-9006-2e666f5b9576-proxy-tls\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598657 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/199ce9b1-1552-4cd4-a84f-67af5391ab73-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-267cw\" (UID: \"199ce9b1-1552-4cd4-a84f-67af5391ab73\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598733 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-865jv\" (UniqueName: \"kubernetes.io/projected/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-kube-api-access-865jv\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598775 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598808 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c152df8-1048-49bb-b981-4853855487d3-srv-cert\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598841 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-default-certificate\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598874 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz8jj\" (UniqueName: \"kubernetes.io/projected/12453f09-2b49-47af-9f73-8c18c93a1812-kube-api-access-qz8jj\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5zmm\" (UID: \"12453f09-2b49-47af-9f73-8c18c93a1812\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598906 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6e3c7165-5c43-4b86-9428-0b0c2846ef7b-cert\") pod \"ingress-canary-z95st\" (UID: \"6e3c7165-5c43-4b86-9428-0b0c2846ef7b\") " pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598957 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjgbd\" (UniqueName: \"kubernetes.io/projected/199ce9b1-1552-4cd4-a84f-67af5391ab73-kube-api-access-sjgbd\") pod \"cluster-samples-operator-665b6dd947-267cw\" (UID: \"199ce9b1-1552-4cd4-a84f-67af5391ab73\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.598996 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/931ec446-cf7a-4040-ae73-05a6ba6c327f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.599033 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a3602e4-822b-4451-90ea-44eaeed085c5-webhook-cert\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.599958 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600081 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600129 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-registration-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600167 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a3602e4-822b-4451-90ea-44eaeed085c5-apiservice-cert\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600232 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600306 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-node-bootstrap-token\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600345 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-socket-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600433 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600484 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600496 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22vt5\" (UniqueName: \"kubernetes.io/projected/4a3602e4-822b-4451-90ea-44eaeed085c5-kube-api-access-22vt5\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600561 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-csi-data-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600614 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/807a87dc-a461-46d7-b116-ca81d06d88a1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-866mj\" (UID: \"807a87dc-a461-46d7-b116-ca81d06d88a1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600648 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/98c86725-7ab9-4054-9006-2e666f5b9576-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600694 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgmsj\" (UniqueName: \"kubernetes.io/projected/b6394941-7bc5-45db-9404-6fa52d0fb10f-kube-api-access-vgmsj\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600856 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn5wc\" (UniqueName: \"kubernetes.io/projected/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-kube-api-access-nn5wc\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600908 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7s8r\" (UniqueName: \"kubernetes.io/projected/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-kube-api-access-p7s8r\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600938 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljmbs\" (UniqueName: \"kubernetes.io/projected/88c5801d-74cf-48c9-8443-8dcd8a743ff3-kube-api-access-ljmbs\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.600970 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601005 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50ffdc3-0fc9-447f-af03-14b6ce63568d-config\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601036 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c374c787-dd0b-472d-914f-e51d1b63ef89-signing-cabundle\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601071 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601095 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kksm6\" (UniqueName: \"kubernetes.io/projected/d3f43d0d-188d-4a54-9800-dd480764d67d-kube-api-access-kksm6\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601126 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c374c787-dd0b-472d-914f-e51d1b63ef89-signing-key\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601157 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601201 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-dir\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601235 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/931ec446-cf7a-4040-ae73-05a6ba6c327f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601311 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6lgr\" (UniqueName: \"kubernetes.io/projected/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-kube-api-access-j6lgr\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601339 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjggg\" (UniqueName: \"kubernetes.io/projected/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-kube-api-access-hjggg\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601366 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/965617c1-858f-4aa3-8c89-185f45434afd-trusted-ca\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601412 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6394941-7bc5-45db-9404-6fa52d0fb10f-secret-volume\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601455 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f43d0d-188d-4a54-9800-dd480764d67d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601482 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-srv-cert\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601510 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601535 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-proxy-tls\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601573 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b4710bf-98cc-4ac1-9356-da1249819e88-service-ca-bundle\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601596 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22j5p\" (UniqueName: \"kubernetes.io/projected/9c4347d9-37c2-4c1e-806b-1c2927c1638b-kube-api-access-22j5p\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601636 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3f43d0d-188d-4a54-9800-dd480764d67d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601663 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxbtt\" (UniqueName: \"kubernetes.io/projected/b87bc33f-e3e7-420d-98de-e648fc610aec-kube-api-access-sxbtt\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601686 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmrm9\" (UniqueName: \"kubernetes.io/projected/2c152df8-1048-49bb-b981-4853855487d3-kube-api-access-lmrm9\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601711 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b87bc33f-e3e7-420d-98de-e648fc610aec-config\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601741 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-plugins-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601772 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjgtj\" (UniqueName: \"kubernetes.io/projected/7b4710bf-98cc-4ac1-9356-da1249819e88-kube-api-access-qjgtj\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601796 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c77h\" (UniqueName: \"kubernetes.io/projected/c374c787-dd0b-472d-914f-e51d1b63ef89-kube-api-access-2c77h\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601819 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601848 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601879 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c152df8-1048-49bb-b981-4853855487d3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601915 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/88c5801d-74cf-48c9-8443-8dcd8a743ff3-metrics-tls\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601912 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/199ce9b1-1552-4cd4-a84f-67af5391ab73-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-267cw\" (UID: \"199ce9b1-1552-4cd4-a84f-67af5391ab73\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601941 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b87bc33f-e3e7-420d-98de-e648fc610aec-serving-cert\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602014 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602085 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/98c86725-7ab9-4054-9006-2e666f5b9576-images\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602122 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-policies\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602149 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-metrics-certs\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602182 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602209 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50ffdc3-0fc9-447f-af03-14b6ce63568d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602264 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/965617c1-858f-4aa3-8c89-185f45434afd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4235e5f-4b25-4e97-b13d-57fcde51d267-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hktj8\" (UID: \"a4235e5f-4b25-4e97-b13d-57fcde51d267\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602312 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-mountpoint-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602334 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931ec446-cf7a-4040-ae73-05a6ba6c327f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602362 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9wpd\" (UniqueName: \"kubernetes.io/projected/807a87dc-a461-46d7-b116-ca81d06d88a1-kube-api-access-s9wpd\") pod \"package-server-manager-789f6589d5-866mj\" (UID: \"807a87dc-a461-46d7-b116-ca81d06d88a1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602401 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d50ffdc3-0fc9-447f-af03-14b6ce63568d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602428 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/965617c1-858f-4aa3-8c89-185f45434afd-metrics-tls\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602463 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7bwb\" (UniqueName: \"kubernetes.io/projected/98c86725-7ab9-4054-9006-2e666f5b9576-kube-api-access-k7bwb\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602504 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-certs\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602528 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6394941-7bc5-45db-9404-6fa52d0fb10f-config-volume\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602562 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602592 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4a3602e4-822b-4451-90ea-44eaeed085c5-tmpfs\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602615 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9g5d\" (UniqueName: \"kubernetes.io/projected/6e3c7165-5c43-4b86-9428-0b0c2846ef7b-kube-api-access-k9g5d\") pod \"ingress-canary-z95st\" (UID: \"6e3c7165-5c43-4b86-9428-0b0c2846ef7b\") " pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602647 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cq8x\" (UniqueName: \"kubernetes.io/projected/965617c1-858f-4aa3-8c89-185f45434afd-kube-api-access-6cq8x\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602672 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bfpv\" (UniqueName: \"kubernetes.io/projected/7030a932-f4cf-4ef3-9db5-9e3a7da7a547-kube-api-access-9bfpv\") pod \"migrator-59844c95c7-25hcn\" (UID: \"7030a932-f4cf-4ef3-9db5-9e3a7da7a547\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602708 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602830 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/931ec446-cf7a-4040-ae73-05a6ba6c327f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.602927 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50ffdc3-0fc9-447f-af03-14b6ce63568d-config\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.603935 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b4710bf-98cc-4ac1-9356-da1249819e88-service-ca-bundle\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.605072 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c152df8-1048-49bb-b981-4853855487d3-srv-cert\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.605166 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/98c86725-7ab9-4054-9006-2e666f5b9576-proxy-tls\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.605229 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a3602e4-822b-4451-90ea-44eaeed085c5-webhook-cert\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.605270 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-socket-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.605312 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-dir\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.605317 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-csi-data-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.605439 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12453f09-2b49-47af-9f73-8c18c93a1812-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5zmm\" (UID: \"12453f09-2b49-47af-9f73-8c18c93a1812\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.606038 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/98c86725-7ab9-4054-9006-2e666f5b9576-images\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.601505 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-registration-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.606182 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f43d0d-188d-4a54-9800-dd480764d67d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.606554 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.106540024 +0000 UTC m=+140.754329356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.606683 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4a3602e4-822b-4451-90ea-44eaeed085c5-tmpfs\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.606744 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-plugins-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.606910 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3f43d0d-188d-4a54-9800-dd480764d67d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.606973 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.607859 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-default-certificate\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.608099 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a3602e4-822b-4451-90ea-44eaeed085c5-apiservice-cert\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.608404 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-proxy-tls\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.608675 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9c4347d9-37c2-4c1e-806b-1c2927c1638b-mountpoint-dir\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.609321 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/98c86725-7ab9-4054-9006-2e666f5b9576-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.609170 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-stats-auth\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.608727 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/965617c1-858f-4aa3-8c89-185f45434afd-trusted-ca\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.609422 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931ec446-cf7a-4040-ae73-05a6ba6c327f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.610186 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.610362 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6394941-7bc5-45db-9404-6fa52d0fb10f-secret-volume\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.610607 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b4710bf-98cc-4ac1-9356-da1249819e88-metrics-certs\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.611844 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50ffdc3-0fc9-447f-af03-14b6ce63568d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.611866 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-srv-cert\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.612317 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4235e5f-4b25-4e97-b13d-57fcde51d267-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hktj8\" (UID: \"a4235e5f-4b25-4e97-b13d-57fcde51d267\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.612600 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c152df8-1048-49bb-b981-4853855487d3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.614781 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/965617c1-858f-4aa3-8c89-185f45434afd-metrics-tls\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.621642 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.631905 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.638605 4741 request.go:700] Waited for 1.004364812s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dmarketplace-trusted-ca&limit=500&resourceVersion=0 Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.645428 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.652271 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.660228 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.681602 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.690795 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/807a87dc-a461-46d7-b116-ca81d06d88a1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-866mj\" (UID: \"807a87dc-a461-46d7-b116-ca81d06d88a1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.700302 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.703919 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.704056 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.204039904 +0000 UTC m=+140.851829236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.704383 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.704709 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.204701736 +0000 UTC m=+140.852491068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.721438 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.725520 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b87bc33f-e3e7-420d-98de-e648fc610aec-serving-cert\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.739630 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.747637 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b87bc33f-e3e7-420d-98de-e648fc610aec-config\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.759581 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.779683 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.801119 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.805293 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.805482 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.305463542 +0000 UTC m=+140.953252884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.805696 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.806054 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.3060439 +0000 UTC m=+140.953833242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.811332 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c374c787-dd0b-472d-914f-e51d1b63ef89-signing-key\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.821276 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.840672 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.844667 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c374c787-dd0b-472d-914f-e51d1b63ef89-signing-cabundle\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.859759 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.881162 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.886693 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.900278 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.903841 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.906997 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.907148 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.407121746 +0000 UTC m=+141.054911078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.907749 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:38 crc kubenswrapper[4741]: E0929 19:11:38.908285 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.408231481 +0000 UTC m=+141.056020943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.927855 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.940033 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.942225 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.947154 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.960045 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.970788 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.981017 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.987948 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:38 crc kubenswrapper[4741]: I0929 19:11:38.999830 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.009807 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.010053 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.51002144 +0000 UTC m=+141.157810812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.011542 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.012081 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.512053485 +0000 UTC m=+141.159842917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.021343 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.035875 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.057135 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.060882 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.068325 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-policies\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.070288 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.080223 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.088342 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.100626 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.106455 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.113095 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.113279 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.613251695 +0000 UTC m=+141.261041037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.113776 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.114173 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.614162835 +0000 UTC m=+141.261952177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.128091 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.136931 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.140544 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.160742 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.179757 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.200426 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.209687 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6394941-7bc5-45db-9404-6fa52d0fb10f-config-volume\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.214750 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.215011 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.714974352 +0000 UTC m=+141.362763694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.215689 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.216181 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.716155959 +0000 UTC m=+141.363945331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.219575 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.229711 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88c5801d-74cf-48c9-8443-8dcd8a743ff3-config-volume\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.241338 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.260142 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.272495 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/88c5801d-74cf-48c9-8443-8dcd8a743ff3-metrics-tls\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.279274 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.300472 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.316804 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.317030 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.816996887 +0000 UTC m=+141.464786259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.317886 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.320848 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.82081977 +0000 UTC m=+141.468609142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.320949 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.340570 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.354458 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6e3c7165-5c43-4b86-9428-0b0c2846ef7b-cert\") pod \"ingress-canary-z95st\" (UID: \"6e3c7165-5c43-4b86-9428-0b0c2846ef7b\") " pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.360947 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.406095 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgxmt\" (UniqueName: \"kubernetes.io/projected/013d0fb9-848e-4156-a67f-498daa95fc6f-kube-api-access-qgxmt\") pod \"openshift-controller-manager-operator-756b6f6bc6-jjbl5\" (UID: \"013d0fb9-848e-4156-a67f-498daa95fc6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.419094 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.419821 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.919780249 +0000 UTC m=+141.567569761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.419962 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.420740 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:39.920715388 +0000 UTC m=+141.568504760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.430348 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b277\" (UniqueName: \"kubernetes.io/projected/8bd18131-53d9-4d13-a85b-41366f624859-kube-api-access-8b277\") pod \"machine-approver-56656f9798-p4mnh\" (UID: \"8bd18131-53d9-4d13-a85b-41366f624859\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.448151 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8gzx\" (UniqueName: \"kubernetes.io/projected/28c53633-9e71-4d02-859e-0cdca912a924-kube-api-access-h8gzx\") pod \"etcd-operator-b45778765-dgp6j\" (UID: \"28c53633-9e71-4d02-859e-0cdca912a924\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.468965 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qjgl\" (UniqueName: \"kubernetes.io/projected/a1ff7353-843d-41bf-86eb-96a315f31bc3-kube-api-access-9qjgl\") pod \"machine-api-operator-5694c8668f-nrp5d\" (UID: \"a1ff7353-843d-41bf-86eb-96a315f31bc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.480678 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.483327 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kws77\" (UniqueName: \"kubernetes.io/projected/c71ee9d0-0827-4cac-aa2f-a6835259d37b-kube-api-access-kws77\") pod \"route-controller-manager-6576b87f9c-4ttw6\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.501033 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.511639 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-node-bootstrap-token\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.521541 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.521884 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.522112 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.022072893 +0000 UTC m=+141.669862255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.522333 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.522908 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.022881939 +0000 UTC m=+141.670671271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.535567 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-certs\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.536324 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.541253 4741 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.560487 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.581540 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.596151 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.616791 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.624092 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.624605 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.124575895 +0000 UTC m=+141.772365227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.624970 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.625553 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.125545805 +0000 UTC m=+141.773335137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.640961 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61b6a0fd-fa98-4401-8a8f-96f0415124e7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.658180 4741 request.go:700] Waited for 1.161494817s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.660167 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-bound-sa-token\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.681628 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.683077 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c26ln\" (UniqueName: \"kubernetes.io/projected/8036b4af-58a3-4e55-9643-4832021a71c5-kube-api-access-c26ln\") pod \"apiserver-76f77b778f-h8d49\" (UID: \"8036b4af-58a3-4e55-9643-4832021a71c5\") " pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.702608 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr9sb\" (UniqueName: \"kubernetes.io/projected/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-kube-api-access-qr9sb\") pod \"console-f9d7485db-jd7m4\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.712275 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5q8m\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-kube-api-access-p5q8m\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.726457 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.727046 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.227029114 +0000 UTC m=+141.874818446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.737578 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmbc\" (UniqueName: \"kubernetes.io/projected/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-kube-api-access-vdmbc\") pod \"controller-manager-879f6c89f-27cn9\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.737841 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.760221 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmr9q\" (UniqueName: \"kubernetes.io/projected/5e2aeccd-cc8d-47ff-801c-8ee9b243415a-kube-api-access-wmr9q\") pod \"apiserver-7bbb656c7d-dz9j9\" (UID: \"5e2aeccd-cc8d-47ff-801c-8ee9b243415a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.776338 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgsk\" (UniqueName: \"kubernetes.io/projected/acdf5b58-ab2f-4526-893c-b9685a63e8a5-kube-api-access-plgsk\") pod \"authentication-operator-69f744f599-pxhts\" (UID: \"acdf5b58-ab2f-4526-893c-b9685a63e8a5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.789585 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.804005 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5jc9\" (UniqueName: \"kubernetes.io/projected/9336afb7-78e3-4ad1-8a48-8401f628d060-kube-api-access-m5jc9\") pod \"openshift-config-operator-7777fb866f-xjn62\" (UID: \"9336afb7-78e3-4ad1-8a48-8401f628d060\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.805946 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.818569 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0efc688-e298-42c1-94c6-c56049a7ec9b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qldrx\" (UID: \"e0efc688-e298-42c1-94c6-c56049a7ec9b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.828850 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.829207 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.329191576 +0000 UTC m=+141.976980908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.835708 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wccqb\" (UniqueName: \"kubernetes.io/projected/5e2985ed-ad44-4d45-9b9b-ad1438579635-kube-api-access-wccqb\") pod \"console-operator-58897d9998-4r5pc\" (UID: \"5e2985ed-ad44-4d45-9b9b-ad1438579635\") " pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.836954 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6"] Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.845658 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" event={"ID":"8bd18131-53d9-4d13-a85b-41366f624859","Type":"ContainerStarted","Data":"599b80131fedbfeafa516c0272b98d90e37bd562c17401d0104e732ea5e3bc3c"} Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.851669 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.856750 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd5l9\" (UniqueName: \"kubernetes.io/projected/3168a724-8ec2-46b6-b151-38e733138a95-kube-api-access-rd5l9\") pod \"openshift-apiserver-operator-796bbdcf4f-99tq6\" (UID: \"3168a724-8ec2-46b6-b151-38e733138a95\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.858322 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.874438 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvskl\" (UniqueName: \"kubernetes.io/projected/68bd7878-4ed3-42e6-859b-9763ebdf8435-kube-api-access-qvskl\") pod \"dns-operator-744455d44c-jtclz\" (UID: \"68bd7878-4ed3-42e6-859b-9763ebdf8435\") " pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.916577 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dgp6j"] Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.917380 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clcjk\" (UniqueName: \"kubernetes.io/projected/bd769734-8dc9-445d-b482-5e8ef9357a0d-kube-api-access-clcjk\") pod \"downloads-7954f5f757-8xdlm\" (UID: \"bd769734-8dc9-445d-b482-5e8ef9357a0d\") " pod="openshift-console/downloads-7954f5f757-8xdlm" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.927936 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.929441 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.930058 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.430019113 +0000 UTC m=+142.077808445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.930436 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:39 crc kubenswrapper[4741]: E0929 19:11:39.930983 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.430970553 +0000 UTC m=+142.078759885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.934514 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mndq4\" (UniqueName: \"kubernetes.io/projected/61b6a0fd-fa98-4401-8a8f-96f0415124e7-kube-api-access-mndq4\") pod \"cluster-image-registry-operator-dc59b4c8b-lfvrr\" (UID: \"61b6a0fd-fa98-4401-8a8f-96f0415124e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.940586 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6m6\" (UniqueName: \"kubernetes.io/projected/a4235e5f-4b25-4e97-b13d-57fcde51d267-kube-api-access-7c6m6\") pod \"multus-admission-controller-857f4d67dd-hktj8\" (UID: \"a4235e5f-4b25-4e97-b13d-57fcde51d267\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.953528 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-865jv\" (UniqueName: \"kubernetes.io/projected/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-kube-api-access-865jv\") pod \"marketplace-operator-79b997595-fjfzz\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.963055 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.970887 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.990660 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nrp5d"] Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.991405 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" Sep 29 19:11:39 crc kubenswrapper[4741]: I0929 19:11:39.992148 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjgbd\" (UniqueName: \"kubernetes.io/projected/199ce9b1-1552-4cd4-a84f-67af5391ab73-kube-api-access-sjgbd\") pod \"cluster-samples-operator-665b6dd947-267cw\" (UID: \"199ce9b1-1552-4cd4-a84f-67af5391ab73\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:39.999085 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz8jj\" (UniqueName: \"kubernetes.io/projected/12453f09-2b49-47af-9f73-8c18c93a1812-kube-api-access-qz8jj\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5zmm\" (UID: \"12453f09-2b49-47af-9f73-8c18c93a1812\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.021314 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgmsj\" (UniqueName: \"kubernetes.io/projected/b6394941-7bc5-45db-9404-6fa52d0fb10f-kube-api-access-vgmsj\") pod \"collect-profiles-29319540-qf6zk\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.034291 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.035398 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.035838 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.535825321 +0000 UTC m=+142.183614653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.048499 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.048545 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn5wc\" (UniqueName: \"kubernetes.io/projected/b3d27f2b-f795-43b9-bdd5-93a79ccb4d86-kube-api-access-nn5wc\") pod \"machine-config-controller-84d6567774-2j5qb\" (UID: \"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.057263 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7s8r\" (UniqueName: \"kubernetes.io/projected/c591c5b8-b43c-46e0-ad4e-fcc71112bd43-kube-api-access-p7s8r\") pod \"machine-config-server-wn992\" (UID: \"c591c5b8-b43c-46e0-ad4e-fcc71112bd43\") " pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.064725 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27cn9"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.077434 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljmbs\" (UniqueName: \"kubernetes.io/projected/88c5801d-74cf-48c9-8443-8dcd8a743ff3-kube-api-access-ljmbs\") pod \"dns-default-f99d7\" (UID: \"88c5801d-74cf-48c9-8443-8dcd8a743ff3\") " pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:40 crc kubenswrapper[4741]: W0929 19:11:40.095754 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff8d9dff_2997_4c1c_9f8a_9f051247d11d.slice/crio-15bc0bb0e855d419fbd1ceca12a79dde385ce53477a3a44f9b37fce057929ea8 WatchSource:0}: Error finding container 15bc0bb0e855d419fbd1ceca12a79dde385ce53477a3a44f9b37fce057929ea8: Status 404 returned error can't find the container with id 15bc0bb0e855d419fbd1ceca12a79dde385ce53477a3a44f9b37fce057929ea8 Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.098171 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.103097 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kksm6\" (UniqueName: \"kubernetes.io/projected/d3f43d0d-188d-4a54-9800-dd480764d67d-kube-api-access-kksm6\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzv59\" (UID: \"d3f43d0d-188d-4a54-9800-dd480764d67d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.121485 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.123269 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xjn62"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.128737 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.130441 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22j5p\" (UniqueName: \"kubernetes.io/projected/9c4347d9-37c2-4c1e-806b-1c2927c1638b-kube-api-access-22j5p\") pod \"csi-hostpathplugin-c6qb4\" (UID: \"9c4347d9-37c2-4c1e-806b-1c2927c1638b\") " pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.138138 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.139733 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.639718207 +0000 UTC m=+142.287507539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.141091 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22vt5\" (UniqueName: \"kubernetes.io/projected/4a3602e4-822b-4451-90ea-44eaeed085c5-kube-api-access-22vt5\") pod \"packageserver-d55dfcdfc-mllcn\" (UID: \"4a3602e4-822b-4451-90ea-44eaeed085c5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.144078 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8xdlm" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.154360 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9wpd\" (UniqueName: \"kubernetes.io/projected/807a87dc-a461-46d7-b116-ca81d06d88a1-kube-api-access-s9wpd\") pod \"package-server-manager-789f6589d5-866mj\" (UID: \"807a87dc-a461-46d7-b116-ca81d06d88a1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.174659 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.176323 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjggg\" (UniqueName: \"kubernetes.io/projected/7e4f9661-755d-4efa-b3ce-91c515e0b2a9-kube-api-access-hjggg\") pod \"catalog-operator-68c6474976-d674b\" (UID: \"7e4f9661-755d-4efa-b3ce-91c515e0b2a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.183858 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.208183 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h8d49"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.218835 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6lgr\" (UniqueName: \"kubernetes.io/projected/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-kube-api-access-j6lgr\") pod \"oauth-openshift-558db77b4-bzlbd\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.222504 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.223623 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cq8x\" (UniqueName: \"kubernetes.io/projected/965617c1-858f-4aa3-8c89-185f45434afd-kube-api-access-6cq8x\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.231904 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.238775 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fjfzz"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.239129 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.239505 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.239926 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.739910294 +0000 UTC m=+142.387699616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.243235 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjgtj\" (UniqueName: \"kubernetes.io/projected/7b4710bf-98cc-4ac1-9356-da1249819e88-kube-api-access-qjgtj\") pod \"router-default-5444994796-ww5bb\" (UID: \"7b4710bf-98cc-4ac1-9356-da1249819e88\") " pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.246173 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.254349 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.262572 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.264231 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmrm9\" (UniqueName: \"kubernetes.io/projected/2c152df8-1048-49bb-b981-4853855487d3-kube-api-access-lmrm9\") pod \"olm-operator-6b444d44fb-bdqsf\" (UID: \"2c152df8-1048-49bb-b981-4853855487d3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.264977 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.279729 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.286073 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxbtt\" (UniqueName: \"kubernetes.io/projected/b87bc33f-e3e7-420d-98de-e648fc610aec-kube-api-access-sxbtt\") pod \"service-ca-operator-777779d784-p6ksg\" (UID: \"b87bc33f-e3e7-420d-98de-e648fc610aec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.293044 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-jd7m4"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.294906 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.297755 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7bwb\" (UniqueName: \"kubernetes.io/projected/98c86725-7ab9-4054-9006-2e666f5b9576-kube-api-access-k7bwb\") pod \"machine-config-operator-74547568cd-pcthr\" (UID: \"98c86725-7ab9-4054-9006-2e666f5b9576\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.305113 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.310452 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.317144 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.323202 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9g5d\" (UniqueName: \"kubernetes.io/projected/6e3c7165-5c43-4b86-9428-0b0c2846ef7b-kube-api-access-k9g5d\") pod \"ingress-canary-z95st\" (UID: \"6e3c7165-5c43-4b86-9428-0b0c2846ef7b\") " pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.323406 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wn992" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.333071 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bfpv\" (UniqueName: \"kubernetes.io/projected/7030a932-f4cf-4ef3-9db5-9e3a7da7a547-kube-api-access-9bfpv\") pod \"migrator-59844c95c7-25hcn\" (UID: \"7030a932-f4cf-4ef3-9db5-9e3a7da7a547\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.340498 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.340816 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.840805594 +0000 UTC m=+142.488594926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.354881 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.367047 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d50ffdc3-0fc9-447f-af03-14b6ce63568d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9q66v\" (UID: \"d50ffdc3-0fc9-447f-af03-14b6ce63568d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.369820 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jtclz"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.387141 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c77h\" (UniqueName: \"kubernetes.io/projected/c374c787-dd0b-472d-914f-e51d1b63ef89-kube-api-access-2c77h\") pod \"service-ca-9c57cc56f-q5rwg\" (UID: \"c374c787-dd0b-472d-914f-e51d1b63ef89\") " pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.405946 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/931ec446-cf7a-4040-ae73-05a6ba6c327f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ftq7r\" (UID: \"931ec446-cf7a-4040-ae73-05a6ba6c327f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.406467 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pxhts"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.417453 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/965617c1-858f-4aa3-8c89-185f45434afd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2tdpd\" (UID: \"965617c1-858f-4aa3-8c89-185f45434afd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.442646 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.443868 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.444171 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:40.944157053 +0000 UTC m=+142.591946385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.468548 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" Sep 29 19:11:40 crc kubenswrapper[4741]: W0929 19:11:40.476253 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3168a724_8ec2_46b6_b151_38e733138a95.slice/crio-bf9440a324a3cdd08f15a2e420c574efbf66b7a7366586a2237e79dca6b6f4a9 WatchSource:0}: Error finding container bf9440a324a3cdd08f15a2e420c574efbf66b7a7366586a2237e79dca6b6f4a9: Status 404 returned error can't find the container with id bf9440a324a3cdd08f15a2e420c574efbf66b7a7366586a2237e79dca6b6f4a9 Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.479378 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx"] Sep 29 19:11:40 crc kubenswrapper[4741]: W0929 19:11:40.481484 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacdf5b58_ab2f_4526_893c_b9685a63e8a5.slice/crio-da38e4f695d337f3d78db166c802b8ef1b8a2a9267650287b20d9bc75a826d46 WatchSource:0}: Error finding container da38e4f695d337f3d78db166c802b8ef1b8a2a9267650287b20d9bc75a826d46: Status 404 returned error can't find the container with id da38e4f695d337f3d78db166c802b8ef1b8a2a9267650287b20d9bc75a826d46 Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.493621 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.509973 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.511007 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.514665 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.539372 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.548222 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.550827 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.050545179 +0000 UTC m=+142.698334511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.591641 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.613288 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8xdlm"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.615902 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-z95st" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.649692 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.650778 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.150754857 +0000 UTC m=+142.798544209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.669409 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.752968 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.753656 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.25364394 +0000 UTC m=+142.901433272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.858771 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.859516 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.35949744 +0000 UTC m=+143.007286772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.905701 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" event={"ID":"ff8d9dff-2997-4c1c-9f8a-9f051247d11d","Type":"ContainerStarted","Data":"06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.905756 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" event={"ID":"ff8d9dff-2997-4c1c-9f8a-9f051247d11d","Type":"ContainerStarted","Data":"15bc0bb0e855d419fbd1ceca12a79dde385ce53477a3a44f9b37fce057929ea8"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.906774 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.907555 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jd7m4" event={"ID":"932a4f81-a3d7-480f-a3bb-648b2e3bbd31","Type":"ContainerStarted","Data":"c9f466d0c22fea5d859fcb3a534ef26a4d6abec61a5c377ab5aa7a93fca5c263"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.908685 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" event={"ID":"68bd7878-4ed3-42e6-859b-9763ebdf8435","Type":"ContainerStarted","Data":"7f0cb52d46218be78be3c32748b0bbae00a900fa99e64a7ff2504e0d014e9a34"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.909835 4741 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-27cn9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.909880 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" podUID="ff8d9dff-2997-4c1c-9f8a-9f051247d11d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.910070 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" event={"ID":"61b6a0fd-fa98-4401-8a8f-96f0415124e7","Type":"ContainerStarted","Data":"dee3429120bbce15ed94f6afef6d4e596f1f8a48d3aa5aebb023bf0edfff1616"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.912111 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" event={"ID":"e0efc688-e298-42c1-94c6-c56049a7ec9b","Type":"ContainerStarted","Data":"79e52e32a0c6218c00227597e4150cb15d35c1129af26bf4a1c6946b07a0e100"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.918242 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" event={"ID":"013d0fb9-848e-4156-a67f-498daa95fc6f","Type":"ContainerStarted","Data":"289508a64fbdd1ca8564634fba844924bc8af2686e10131190a41484ac6ba857"} Sep 29 19:11:40 crc kubenswrapper[4741]: W0929 19:11:40.918627 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd769734_8dc9_445d_b482_5e8ef9357a0d.slice/crio-df80dc53b72418bcd0e02959197dc9c3a1bc1942295788f56c96a9b849316fc0 WatchSource:0}: Error finding container df80dc53b72418bcd0e02959197dc9c3a1bc1942295788f56c96a9b849316fc0: Status 404 returned error can't find the container with id df80dc53b72418bcd0e02959197dc9c3a1bc1942295788f56c96a9b849316fc0 Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.918944 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" event={"ID":"8036b4af-58a3-4e55-9643-4832021a71c5","Type":"ContainerStarted","Data":"c2310d277954dbc885718f5a33c555394b5013919d5ecf8ba01c74bbfd387cb1"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.919639 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" event={"ID":"5e2aeccd-cc8d-47ff-801c-8ee9b243415a","Type":"ContainerStarted","Data":"d49aa63aa5e7d3a8016d79d289744d3e03699d361c8c9f3908a7cac9b99d1c5c"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.922782 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" event={"ID":"a1ff7353-843d-41bf-86eb-96a315f31bc3","Type":"ContainerStarted","Data":"67747edebf096c736b8a5370b17d1c5eee23e898cb8bc797609beb92ddf92dfe"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.922804 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" event={"ID":"a1ff7353-843d-41bf-86eb-96a315f31bc3","Type":"ContainerStarted","Data":"11ccf491b18d30fcbc3c1542b970b9d5fa60ebc89e927a9fdaa7b07d1e062e52"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.930743 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" event={"ID":"acdf5b58-ab2f-4526-893c-b9685a63e8a5","Type":"ContainerStarted","Data":"da38e4f695d337f3d78db166c802b8ef1b8a2a9267650287b20d9bc75a826d46"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.937863 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" event={"ID":"3168a724-8ec2-46b6-b151-38e733138a95","Type":"ContainerStarted","Data":"bf9440a324a3cdd08f15a2e420c574efbf66b7a7366586a2237e79dca6b6f4a9"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.938965 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" event={"ID":"9336afb7-78e3-4ad1-8a48-8401f628d060","Type":"ContainerStarted","Data":"bc935fcb0666e43f413a8a1bff31b06466028ff15dfb34638b19c0979fa5b422"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.940129 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" event={"ID":"28c53633-9e71-4d02-859e-0cdca912a924","Type":"ContainerStarted","Data":"a7bc03351e0559ce95bf8b8d737ad4c17484731357f7047f46165d8e26f2eaaa"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.940149 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" event={"ID":"28c53633-9e71-4d02-859e-0cdca912a924","Type":"ContainerStarted","Data":"5284085d3934f06ba652eecb11168391e4945160764cb3bf434a2641dcdbd099"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.940839 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" event={"ID":"2d7a0ce6-fda6-4100-801b-a7170a4d5e93","Type":"ContainerStarted","Data":"64633855a131143b04a5b36e65a947c4bccfe8637320edd11453d0cf810e4911"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.951776 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" event={"ID":"8bd18131-53d9-4d13-a85b-41366f624859","Type":"ContainerStarted","Data":"35747c37b1295852a8f77fd7c3f47ff78d0a117fcf2e3f72e9767279a507a7d8"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.951814 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" event={"ID":"8bd18131-53d9-4d13-a85b-41366f624859","Type":"ContainerStarted","Data":"53bae7d5c1ff219f711fe20cff00601a75c7e58734d7c136e32bed54bc8f982b"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.956463 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4r5pc"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.961307 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:40 crc kubenswrapper[4741]: E0929 19:11:40.961625 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.461613169 +0000 UTC m=+143.109402501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.961882 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" event={"ID":"c71ee9d0-0827-4cac-aa2f-a6835259d37b","Type":"ContainerStarted","Data":"78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.961905 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" event={"ID":"c71ee9d0-0827-4cac-aa2f-a6835259d37b","Type":"ContainerStarted","Data":"9ec4ea6ac1589d01dbcdb1d1d1cc2f2da2e02bcd2ececa12b885c21778013644"} Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.962259 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.972084 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw"] Sep 29 19:11:40 crc kubenswrapper[4741]: I0929 19:11:40.987579 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.005614 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.057515 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.060601 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.062658 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.063630 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj"] Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.064526 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.564509803 +0000 UTC m=+143.212299135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.138379 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-f99d7"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.149560 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.174637 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.175566 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.675544619 +0000 UTC m=+143.323333951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.258806 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.288114 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.291110 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.791091358 +0000 UTC m=+143.438880690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.356369 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-c6qb4"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.391933 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.392753 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.892741723 +0000 UTC m=+143.540531055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.393698 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" podStartSLOduration=123.393686854 podStartE2EDuration="2m3.393686854s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:41.388207928 +0000 UTC m=+143.035997260" watchObservedRunningTime="2025-09-29 19:11:41.393686854 +0000 UTC m=+143.041476186" Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.398290 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.495413 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.496025 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:41.996009859 +0000 UTC m=+143.643799191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.541145 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" podStartSLOduration=123.541129878 podStartE2EDuration="2m3.541129878s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:41.537618945 +0000 UTC m=+143.185408277" watchObservedRunningTime="2025-09-29 19:11:41.541129878 +0000 UTC m=+143.188919210" Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.577085 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p4mnh" podStartSLOduration=123.577064652 podStartE2EDuration="2m3.577064652s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:41.574061195 +0000 UTC m=+143.221850547" watchObservedRunningTime="2025-09-29 19:11:41.577064652 +0000 UTC m=+143.224853984" Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.596555 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.596994 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.096974651 +0000 UTC m=+143.744764033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.697494 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.698197 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.198180951 +0000 UTC m=+143.845970283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.698266 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.698593 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.198586424 +0000 UTC m=+143.846375756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.797703 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hktj8"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.803861 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.806248 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.30622087 +0000 UTC m=+143.954010202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.810363 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk"] Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.853173 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dgp6j" podStartSLOduration=123.853158028 podStartE2EDuration="2m3.853158028s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:41.850198803 +0000 UTC m=+143.497988135" watchObservedRunningTime="2025-09-29 19:11:41.853158028 +0000 UTC m=+143.500947360" Sep 29 19:11:41 crc kubenswrapper[4741]: W0929 19:11:41.893824 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4235e5f_4b25_4e97_b13d_57fcde51d267.slice/crio-9f1c775b3676fe0b600e81e826f83f9e06cb3ff59712d200e2834d08dda4c1f5 WatchSource:0}: Error finding container 9f1c775b3676fe0b600e81e826f83f9e06cb3ff59712d200e2834d08dda4c1f5: Status 404 returned error can't find the container with id 9f1c775b3676fe0b600e81e826f83f9e06cb3ff59712d200e2834d08dda4c1f5 Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.912872 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:41 crc kubenswrapper[4741]: E0929 19:11:41.913123 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.413111023 +0000 UTC m=+144.060900355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.991197 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" event={"ID":"013d0fb9-848e-4156-a67f-498daa95fc6f","Type":"ContainerStarted","Data":"8a8cbacdd8e821b214ddf7e4b5f3986f5a435a48e2d2d9c24f4b3fc614448631"} Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.995944 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" event={"ID":"7e4f9661-755d-4efa-b3ce-91c515e0b2a9","Type":"ContainerStarted","Data":"1821cab92eb6f3839c67783c3a646d3fe7a1211c66fbde759472dda39de264aa"} Sep 29 19:11:41 crc kubenswrapper[4741]: I0929 19:11:41.997971 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" event={"ID":"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86","Type":"ContainerStarted","Data":"ad39548d5f199654bb0cba15d505c654de5a0f9f9501cc9af900306ebc7c8e17"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:41.999776 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" event={"ID":"d50ffdc3-0fc9-447f-af03-14b6ce63568d","Type":"ContainerStarted","Data":"f0acbf78b821da13e4c4c34fdcd82207a98b42dcaee1fd63f637092817684c3a"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.005788 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" event={"ID":"68bd7878-4ed3-42e6-859b-9763ebdf8435","Type":"ContainerStarted","Data":"bffc82f9bf6facff4e43c45c3c57f97808d8403f9ca81d4d2f08d1de42baaa6f"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.010009 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" event={"ID":"d3f43d0d-188d-4a54-9800-dd480764d67d","Type":"ContainerStarted","Data":"351ce2d4fef9db3d4d79d7b15f76f78741087e77974deb31c8b9c1490bd1f07f"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.018508 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.019882 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.519860471 +0000 UTC m=+144.167649813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.035716 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.055297 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" event={"ID":"5e2985ed-ad44-4d45-9b9b-ad1438579635","Type":"ContainerStarted","Data":"fa14671aa4281292944b6af622050b383336a9baecccc59ba6ea58b4b5ccc3f5"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.055353 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" event={"ID":"5e2985ed-ad44-4d45-9b9b-ad1438579635","Type":"ContainerStarted","Data":"99c5a6f76ed92f7361b877bbe8fa727b333ff074d07065a7a0ab5f097c9498ed"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.056471 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.068445 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wn992" event={"ID":"c591c5b8-b43c-46e0-ad4e-fcc71112bd43","Type":"ContainerStarted","Data":"b121d40a1ac265598b452402bae63d12cf58620f90d979326cadf99b01087460"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.079556 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.089548 4741 patch_prober.go:28] interesting pod/console-operator-58897d9998-4r5pc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.089599 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" podUID="5e2985ed-ad44-4d45-9b9b-ad1438579635" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.089647 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" event={"ID":"12453f09-2b49-47af-9f73-8c18c93a1812","Type":"ContainerStarted","Data":"c7b8fc5677a7821bd0f6b679b779ca543693d008e7d9247498824c22e6366781"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.091944 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ww5bb" event={"ID":"7b4710bf-98cc-4ac1-9356-da1249819e88","Type":"ContainerStarted","Data":"ca5cf9aef08b71cf50f134de84b4936606af26d8db3ad48caca29961c74eafa3"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.094731 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jd7m4" event={"ID":"932a4f81-a3d7-480f-a3bb-648b2e3bbd31","Type":"ContainerStarted","Data":"fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.097413 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.100644 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzlbd"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.112362 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-f99d7" event={"ID":"88c5801d-74cf-48c9-8443-8dcd8a743ff3","Type":"ContainerStarted","Data":"a52326e7174be29b29718f0504e0b69a02d440c4d5489d1caeaa8719e7c0e11d"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.115453 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.118543 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.121238 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.122717 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.622702643 +0000 UTC m=+144.270491975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.145603 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q5rwg"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.180583 4741 generic.go:334] "Generic (PLEG): container finished" podID="9336afb7-78e3-4ad1-8a48-8401f628d060" containerID="b0f89c2c2eac75a184eb6dad5f5dc3ada3dc27b18a20c77bd85259646cfc3df0" exitCode=0 Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.180928 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" event={"ID":"9336afb7-78e3-4ad1-8a48-8401f628d060","Type":"ContainerDied","Data":"b0f89c2c2eac75a184eb6dad5f5dc3ada3dc27b18a20c77bd85259646cfc3df0"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.198021 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.211367 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-z95st"] Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.222313 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.222952 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.722927771 +0000 UTC m=+144.370717163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: W0929 19:11:42.225372 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98c86725_7ab9_4054_9006_2e666f5b9576.slice/crio-370c85ca8bc24530508095822e37f29f4314547a13784b866684fca2c663c00f WatchSource:0}: Error finding container 370c85ca8bc24530508095822e37f29f4314547a13784b866684fca2c663c00f: Status 404 returned error can't find the container with id 370c85ca8bc24530508095822e37f29f4314547a13784b866684fca2c663c00f Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.242992 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" event={"ID":"b6394941-7bc5-45db-9404-6fa52d0fb10f","Type":"ContainerStarted","Data":"51d0652cb7115bd7d2bc2ac922a38bccac18abe2e5e26efa2530783393291c82"} Sep 29 19:11:42 crc kubenswrapper[4741]: W0929 19:11:42.251099 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb87bc33f_e3e7_420d_98de_e648fc610aec.slice/crio-4716322a956ef845d5462611667b44c906779df0bd572e53ede3bb9579f8bfc6 WatchSource:0}: Error finding container 4716322a956ef845d5462611667b44c906779df0bd572e53ede3bb9579f8bfc6: Status 404 returned error can't find the container with id 4716322a956ef845d5462611667b44c906779df0bd572e53ede3bb9579f8bfc6 Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.251128 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" event={"ID":"a1ff7353-843d-41bf-86eb-96a315f31bc3","Type":"ContainerStarted","Data":"4f5405fb2e0a20df19ede6971bbd1143cfeccd6b954628c55bfbd16cee6ab894"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.276473 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" event={"ID":"61b6a0fd-fa98-4401-8a8f-96f0415124e7","Type":"ContainerStarted","Data":"b51fa117ac312aff29762ef4d5a6745a4e610585e7e59fc4c2af3674155fa3f8"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.311738 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" event={"ID":"199ce9b1-1552-4cd4-a84f-67af5391ab73","Type":"ContainerStarted","Data":"32dabd4fb4c200530c597ca6bdbc1df60a5e257410ee15cae9f8f768767c363a"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.323507 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.324567 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.824556255 +0000 UTC m=+144.472345587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.335735 4741 generic.go:334] "Generic (PLEG): container finished" podID="8036b4af-58a3-4e55-9643-4832021a71c5" containerID="a74920fa62e8ef029ba5fbe6608a62787093cdae63b85d63e51470c9135df67b" exitCode=0 Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.335802 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" event={"ID":"8036b4af-58a3-4e55-9643-4832021a71c5","Type":"ContainerDied","Data":"a74920fa62e8ef029ba5fbe6608a62787093cdae63b85d63e51470c9135df67b"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.397901 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" event={"ID":"acdf5b58-ab2f-4526-893c-b9685a63e8a5","Type":"ContainerStarted","Data":"2346dfca364a823a35dcb828e60e552e8ef27d5c37cbe3a6f8e106cb67407d70"} Sep 29 19:11:42 crc kubenswrapper[4741]: W0929 19:11:42.405202 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7030a932_f4cf_4ef3_9db5_9e3a7da7a547.slice/crio-cff7d073298b28c3699a39e5911260a24437dee99bba9bb02997c69fb699401c WatchSource:0}: Error finding container cff7d073298b28c3699a39e5911260a24437dee99bba9bb02997c69fb699401c: Status 404 returned error can't find the container with id cff7d073298b28c3699a39e5911260a24437dee99bba9bb02997c69fb699401c Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.405850 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" event={"ID":"9c4347d9-37c2-4c1e-806b-1c2927c1638b","Type":"ContainerStarted","Data":"ea0a47f613f0013f681a4c238689ffac0243dd0ac41a0c8240d23cd5a1c8b382"} Sep 29 19:11:42 crc kubenswrapper[4741]: W0929 19:11:42.408803 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e3c7165_5c43_4b86_9428_0b0c2846ef7b.slice/crio-70287d4dd7f02575bdc609d788ecf83e7548948a1a9bef398e2d59b61575cdc6 WatchSource:0}: Error finding container 70287d4dd7f02575bdc609d788ecf83e7548948a1a9bef398e2d59b61575cdc6: Status 404 returned error can't find the container with id 70287d4dd7f02575bdc609d788ecf83e7548948a1a9bef398e2d59b61575cdc6 Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.425222 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.426361 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:42.926343343 +0000 UTC m=+144.574132675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.456450 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" event={"ID":"4a3602e4-822b-4451-90ea-44eaeed085c5","Type":"ContainerStarted","Data":"c5c0d8e9e2fda476d2f51cd78daa5da03aa2596f7631664711ab16c1d73285b6"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.457523 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.470457 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" event={"ID":"3168a724-8ec2-46b6-b151-38e733138a95","Type":"ContainerStarted","Data":"6b4a954efe7ec18a335686fc59c55bc80326c3061162045c286f5f4ea3ff5301"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.492876 4741 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-mllcn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.492936 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" podUID="4a3602e4-822b-4451-90ea-44eaeed085c5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.510055 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxhts" podStartSLOduration=124.510035971 podStartE2EDuration="2m4.510035971s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.491512206 +0000 UTC m=+144.139301538" watchObservedRunningTime="2025-09-29 19:11:42.510035971 +0000 UTC m=+144.157825303" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.519662 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" event={"ID":"2d7a0ce6-fda6-4100-801b-a7170a4d5e93","Type":"ContainerStarted","Data":"68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.520720 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.530257 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8xdlm" event={"ID":"bd769734-8dc9-445d-b482-5e8ef9357a0d","Type":"ContainerStarted","Data":"df80dc53b72418bcd0e02959197dc9c3a1bc1942295788f56c96a9b849316fc0"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.530372 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.531111 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8xdlm" Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.531426 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.031410557 +0000 UTC m=+144.679199889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.541863 4741 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fjfzz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.541920 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" podUID="2d7a0ce6-fda6-4100-801b-a7170a4d5e93" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.544813 4741 patch_prober.go:28] interesting pod/downloads-7954f5f757-8xdlm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.544865 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8xdlm" podUID="bd769734-8dc9-445d-b482-5e8ef9357a0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.564212 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" event={"ID":"a4235e5f-4b25-4e97-b13d-57fcde51d267","Type":"ContainerStarted","Data":"9f1c775b3676fe0b600e81e826f83f9e06cb3ff59712d200e2834d08dda4c1f5"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.579520 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" event={"ID":"807a87dc-a461-46d7-b116-ca81d06d88a1","Type":"ContainerStarted","Data":"5d76645b93947cabd487dc2aec3238156163b6594082fe523349a658dd4b0f03"} Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.597613 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-jd7m4" podStartSLOduration=124.597587462 podStartE2EDuration="2m4.597587462s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.5884723 +0000 UTC m=+144.236261632" watchObservedRunningTime="2025-09-29 19:11:42.597587462 +0000 UTC m=+144.245376794" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.605030 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.610020 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfvrr" podStartSLOduration=124.610003171 podStartE2EDuration="2m4.610003171s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.608626387 +0000 UTC m=+144.256415749" watchObservedRunningTime="2025-09-29 19:11:42.610003171 +0000 UTC m=+144.257792503" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.631325 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.631477 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.13145447 +0000 UTC m=+144.779243802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.631593 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.631971 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.131947385 +0000 UTC m=+144.779736717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.647876 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" podStartSLOduration=124.647854397 podStartE2EDuration="2m4.647854397s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.642271217 +0000 UTC m=+144.290060549" watchObservedRunningTime="2025-09-29 19:11:42.647854397 +0000 UTC m=+144.295643729" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.685917 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jjbl5" podStartSLOduration=124.685887947 podStartE2EDuration="2m4.685887947s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.681856059 +0000 UTC m=+144.329645391" watchObservedRunningTime="2025-09-29 19:11:42.685887947 +0000 UTC m=+144.333677279" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.738229 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-nrp5d" podStartSLOduration=124.738210837 podStartE2EDuration="2m4.738210837s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.737737093 +0000 UTC m=+144.385526425" watchObservedRunningTime="2025-09-29 19:11:42.738210837 +0000 UTC m=+144.386000179" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.739984 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.740224 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.240198192 +0000 UTC m=+144.887987524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.740312 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.740875 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.240859092 +0000 UTC m=+144.888648414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.841016 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.841496 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.341479334 +0000 UTC m=+144.989268666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.865126 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" podStartSLOduration=124.865112743 podStartE2EDuration="2m4.865112743s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.839158149 +0000 UTC m=+144.486947481" watchObservedRunningTime="2025-09-29 19:11:42.865112743 +0000 UTC m=+144.512902075" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.923530 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99tq6" podStartSLOduration=124.923493997 podStartE2EDuration="2m4.923493997s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.885610231 +0000 UTC m=+144.533399563" watchObservedRunningTime="2025-09-29 19:11:42.923493997 +0000 UTC m=+144.571283329" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.923978 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8xdlm" podStartSLOduration=124.923973663 podStartE2EDuration="2m4.923973663s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.920950936 +0000 UTC m=+144.568740278" watchObservedRunningTime="2025-09-29 19:11:42.923973663 +0000 UTC m=+144.571762995" Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.945105 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:42 crc kubenswrapper[4741]: E0929 19:11:42.945484 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.445473873 +0000 UTC m=+145.093263205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:42 crc kubenswrapper[4741]: I0929 19:11:42.958568 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" podStartSLOduration=124.958522583 podStartE2EDuration="2m4.958522583s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:42.956088454 +0000 UTC m=+144.603877796" watchObservedRunningTime="2025-09-29 19:11:42.958522583 +0000 UTC m=+144.606311915" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.050980 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.051350 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.551334383 +0000 UTC m=+145.199123715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.155221 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.155897 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.655882829 +0000 UTC m=+145.303672161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.262705 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.262965 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.762938007 +0000 UTC m=+145.410727339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.263059 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.263433 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.763378932 +0000 UTC m=+145.411168264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.382481 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.382651 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.88262645 +0000 UTC m=+145.530415782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.383039 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.383435 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.883384385 +0000 UTC m=+145.531173727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.483910 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.485617 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:43.985601017 +0000 UTC m=+145.633390349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.586674 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.587006 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.086994693 +0000 UTC m=+145.734784025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.606585 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" event={"ID":"d3f43d0d-188d-4a54-9800-dd480764d67d","Type":"ContainerStarted","Data":"a58a434179906351a58fc61080ea2b3fb05bc943530c3f81ee748ccb00ca9ab3"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.619995 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" event={"ID":"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86","Type":"ContainerStarted","Data":"fbb6fde09c6b57e7585cfd0f4e6ca3ad5bfefd7538ee22e99891d4b18bea170b"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.620035 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" event={"ID":"b3d27f2b-f795-43b9-bdd5-93a79ccb4d86","Type":"ContainerStarted","Data":"8d3e6fdaa5c4fb201e4ceca6cc56916cdd4ec0d6a3a2cc031bda7f6f9a4276b5"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.628710 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8xdlm" event={"ID":"bd769734-8dc9-445d-b482-5e8ef9357a0d","Type":"ContainerStarted","Data":"8dab423627ea8bf59e59735fd58575cfcf095816597724e8a9f78f94d6d2c678"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.630608 4741 patch_prober.go:28] interesting pod/downloads-7954f5f757-8xdlm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.630665 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8xdlm" podUID="bd769734-8dc9-445d-b482-5e8ef9357a0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.651869 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" event={"ID":"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6","Type":"ContainerStarted","Data":"e4bbf68e9561016712f397ac6841b4cfc0062e71b0e1fc66e2b4f579385ea4a8"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.653441 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" event={"ID":"9336afb7-78e3-4ad1-8a48-8401f628d060","Type":"ContainerStarted","Data":"1f231b040fe0be0bf4540bed2b88bf8cbd2129ee376699f059891620b215d720"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.654058 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.658664 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" event={"ID":"c374c787-dd0b-472d-914f-e51d1b63ef89","Type":"ContainerStarted","Data":"1f3b09caf41a7e3b96ef7e6c912f3ca6cb8bcb315fd4cf7fc5af30e6f988e5a4"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.658701 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" event={"ID":"c374c787-dd0b-472d-914f-e51d1b63ef89","Type":"ContainerStarted","Data":"9bfa46a55ce7bc776701f9d21929b9138d09a5a28cf0094c2fa9da927550f446"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.665873 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" event={"ID":"d50ffdc3-0fc9-447f-af03-14b6ce63568d","Type":"ContainerStarted","Data":"72dba2ba7c939bbb9645416ef8516b9aa9b7ff9ef267413c5cd2c0977aac3c46"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.688118 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.688645 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.188630767 +0000 UTC m=+145.836420099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.721456 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" event={"ID":"68bd7878-4ed3-42e6-859b-9763ebdf8435","Type":"ContainerStarted","Data":"89cd2206ed4a67be31591715bc28f402100863d90338718e29a9b315d9b83435"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.739512 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2j5qb" podStartSLOduration=125.73949441 podStartE2EDuration="2m5.73949441s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:43.73668309 +0000 UTC m=+145.384472422" watchObservedRunningTime="2025-09-29 19:11:43.73949441 +0000 UTC m=+145.387283742" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.740909 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzv59" podStartSLOduration=125.740890375 podStartE2EDuration="2m5.740890375s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:43.639044485 +0000 UTC m=+145.286833827" watchObservedRunningTime="2025-09-29 19:11:43.740890375 +0000 UTC m=+145.388679707" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.741660 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" event={"ID":"4a3602e4-822b-4451-90ea-44eaeed085c5","Type":"ContainerStarted","Data":"96fe15d07d070d5e8e34da468aa2e52c0aa741a86bc6729136739beba2f106f8"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.756484 4741 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-mllcn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.756536 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" podUID="4a3602e4-822b-4451-90ea-44eaeed085c5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.796499 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.799702 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.299685473 +0000 UTC m=+145.947474805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.804681 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" event={"ID":"12453f09-2b49-47af-9f73-8c18c93a1812","Type":"ContainerStarted","Data":"368d5b5eaa96b56fd4e994e69965d5e114c741541915c497319012ab0d65b4fb"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.866553 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-z95st" event={"ID":"6e3c7165-5c43-4b86-9428-0b0c2846ef7b","Type":"ContainerStarted","Data":"65edf9038938b108a61a151d2a5d55491ab9d43fdc5239d11fe1b7dd1364c574"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.866596 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-z95st" event={"ID":"6e3c7165-5c43-4b86-9428-0b0c2846ef7b","Type":"ContainerStarted","Data":"70287d4dd7f02575bdc609d788ecf83e7548948a1a9bef398e2d59b61575cdc6"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.898783 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" event={"ID":"199ce9b1-1552-4cd4-a84f-67af5391ab73","Type":"ContainerStarted","Data":"90efc9c28f8627daa0b1abbb94b4754063d2e37099cc60a2cef231be9e8863ec"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.899058 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" event={"ID":"199ce9b1-1552-4cd4-a84f-67af5391ab73","Type":"ContainerStarted","Data":"452bd708c6b558c0a442a436e1db5508e6025149019a584027578ea6d548b237"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.899080 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:43 crc kubenswrapper[4741]: E0929 19:11:43.899768 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.399753137 +0000 UTC m=+146.047542459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.947701 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" event={"ID":"98c86725-7ab9-4054-9006-2e666f5b9576","Type":"ContainerStarted","Data":"69493dfa80279af190aef8e5ad634bd34cb45e4562337484468666f13b45a97f"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.947991 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" event={"ID":"98c86725-7ab9-4054-9006-2e666f5b9576","Type":"ContainerStarted","Data":"370c85ca8bc24530508095822e37f29f4314547a13784b866684fca2c663c00f"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.966519 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-f99d7" event={"ID":"88c5801d-74cf-48c9-8443-8dcd8a743ff3","Type":"ContainerStarted","Data":"fbeb35085c4cae9143eb9e96b13b392cb37ec53141f47a0a100cf25148633443"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.967489 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.968689 4741 generic.go:334] "Generic (PLEG): container finished" podID="5e2aeccd-cc8d-47ff-801c-8ee9b243415a" containerID="a0d88eea291d380a41e6ea64352f86ecefd39df9d6b1f9ac4aea7d0c8c7df256" exitCode=0 Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.968732 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" event={"ID":"5e2aeccd-cc8d-47ff-801c-8ee9b243415a","Type":"ContainerDied","Data":"a0d88eea291d380a41e6ea64352f86ecefd39df9d6b1f9ac4aea7d0c8c7df256"} Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.977151 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" podStartSLOduration=125.977134801 podStartE2EDuration="2m5.977134801s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:43.976664976 +0000 UTC m=+145.624454308" watchObservedRunningTime="2025-09-29 19:11:43.977134801 +0000 UTC m=+145.624924133" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.977754 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q66v" podStartSLOduration=125.977749271 podStartE2EDuration="2m5.977749271s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:43.860671291 +0000 UTC m=+145.508460623" watchObservedRunningTime="2025-09-29 19:11:43.977749271 +0000 UTC m=+145.625538593" Sep 29 19:11:43 crc kubenswrapper[4741]: I0929 19:11:43.989056 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" event={"ID":"a4235e5f-4b25-4e97-b13d-57fcde51d267","Type":"ContainerStarted","Data":"7ab4cb2fbb4358563d8cda10e0a37d5dbad2099b1218c54cda2b982970033108"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.005681 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.006974 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.506960329 +0000 UTC m=+146.154749661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.015619 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wn992" event={"ID":"c591c5b8-b43c-46e0-ad4e-fcc71112bd43","Type":"ContainerStarted","Data":"911d6beb5c943a2de43e2632668e78dea2d7dd1e84afee47fa251dee55d0ef85"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.039408 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jtclz" podStartSLOduration=126.03936715 podStartE2EDuration="2m6.03936715s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.037961924 +0000 UTC m=+145.685751256" watchObservedRunningTime="2025-09-29 19:11:44.03936715 +0000 UTC m=+145.687156482" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.057893 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" event={"ID":"807a87dc-a461-46d7-b116-ca81d06d88a1","Type":"ContainerStarted","Data":"b0c40d1dcc8159f531cc99fa187e13bfb6100b74a43f1e24590c14ef985cbfd6"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.057933 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" event={"ID":"807a87dc-a461-46d7-b116-ca81d06d88a1","Type":"ContainerStarted","Data":"25c1d9a52ed44de9efeb529c01190da185d40ad2f4e9b6ae851b38b7d327415b"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.058789 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.060852 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" event={"ID":"931ec446-cf7a-4040-ae73-05a6ba6c327f","Type":"ContainerStarted","Data":"a3780da773e9d59b7a6dd0f89bd0e0db9e3e33707f04ab11bbe02633a9edf9ab"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.060873 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" event={"ID":"931ec446-cf7a-4040-ae73-05a6ba6c327f","Type":"ContainerStarted","Data":"7d3d7d5301478590df5e9fe931354077852f412319979356f5fa0f7029ec32e5"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.072821 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" event={"ID":"b6394941-7bc5-45db-9404-6fa52d0fb10f","Type":"ContainerStarted","Data":"292180dc73a15c3eab4c0cf1929f8f62499d6f33cd171557c9522193af31e653"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.108176 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.109479 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.609462661 +0000 UTC m=+146.257251993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.115684 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" event={"ID":"8036b4af-58a3-4e55-9643-4832021a71c5","Type":"ContainerStarted","Data":"dc8fc35eb18f6e089d64a9fa47a7999275dfdd7d51abe913fbe20140b80b2439"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.124794 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-q5rwg" podStartSLOduration=125.124780362 podStartE2EDuration="2m5.124780362s" podCreationTimestamp="2025-09-29 19:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.088641222 +0000 UTC m=+145.736430554" watchObservedRunningTime="2025-09-29 19:11:44.124780362 +0000 UTC m=+145.772569694" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.145184 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" event={"ID":"965617c1-858f-4aa3-8c89-185f45434afd","Type":"ContainerStarted","Data":"842d9834022d98f748bd4a64ccd3c364e831c2ae4873c5b76d1105c4f46a8966"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.145230 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" event={"ID":"965617c1-858f-4aa3-8c89-185f45434afd","Type":"ContainerStarted","Data":"957dcfd2e9144582eb7be7e00a17baad679d94ace8983c33f86b94a2f298d1e7"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.150134 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" event={"ID":"7e4f9661-755d-4efa-b3ce-91c515e0b2a9","Type":"ContainerStarted","Data":"28f97e772e2bb46be9400b48d2d159256e666da91c061035ad0fab1887170384"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.151194 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.156814 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-f99d7" podStartSLOduration=7.156799771 podStartE2EDuration="7.156799771s" podCreationTimestamp="2025-09-29 19:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.154216478 +0000 UTC m=+145.802005810" watchObservedRunningTime="2025-09-29 19:11:44.156799771 +0000 UTC m=+145.804589103" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.167922 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ww5bb" event={"ID":"7b4710bf-98cc-4ac1-9356-da1249819e88","Type":"ContainerStarted","Data":"fd586183d1b7a3d0f7bdd666ce390dabb8d355035f8ec6f0e358cd8fa8e27407"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.171629 4741 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-d674b container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.171670 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" podUID="7e4f9661-755d-4efa-b3ce-91c515e0b2a9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.181741 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" event={"ID":"b87bc33f-e3e7-420d-98de-e648fc610aec","Type":"ContainerStarted","Data":"338a14ac0400f4dcc0d6793e36dca02aadb9ebda23c7c554e9eb93b4ac4a455c"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.181786 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" event={"ID":"b87bc33f-e3e7-420d-98de-e648fc610aec","Type":"ContainerStarted","Data":"4716322a956ef845d5462611667b44c906779df0bd572e53ede3bb9579f8bfc6"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.201360 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" event={"ID":"2c152df8-1048-49bb-b981-4853855487d3","Type":"ContainerStarted","Data":"ac925b0d04456df832255c34724d01cc9e9993e988f2a0b8d0e9a93a70ebfad8"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.201433 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" event={"ID":"2c152df8-1048-49bb-b981-4853855487d3","Type":"ContainerStarted","Data":"cc3172c7698434e59f57919d275d12498a2cc372400ed310ceb1e714c677b680"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.202250 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.209506 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.211051 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.711040472 +0000 UTC m=+146.358829794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.214793 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5zmm" podStartSLOduration=126.214775062 podStartE2EDuration="2m6.214775062s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.21160745 +0000 UTC m=+145.859396782" watchObservedRunningTime="2025-09-29 19:11:44.214775062 +0000 UTC m=+145.862564394" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.227728 4741 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bdqsf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.227796 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" podUID="2c152df8-1048-49bb-b981-4853855487d3" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.249095 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.260219 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:44 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:44 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:44 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.260263 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.260262 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" event={"ID":"e0efc688-e298-42c1-94c6-c56049a7ec9b","Type":"ContainerStarted","Data":"43b011d13f10b36c4a420ae6fd5b14106b93779450675e7fee32f2bd032d35bc"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.266275 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" event={"ID":"7030a932-f4cf-4ef3-9db5-9e3a7da7a547","Type":"ContainerStarted","Data":"cff7d073298b28c3699a39e5911260a24437dee99bba9bb02997c69fb699401c"} Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.284572 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.287186 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ftq7r" podStartSLOduration=126.287158066 podStartE2EDuration="2m6.287158066s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.249513808 +0000 UTC m=+145.897303140" watchObservedRunningTime="2025-09-29 19:11:44.287158066 +0000 UTC m=+145.934947398" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.311072 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.311621 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.811604371 +0000 UTC m=+146.459393703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.322184 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-z95st" podStartSLOduration=7.322166921 podStartE2EDuration="7.322166921s" podCreationTimestamp="2025-09-29 19:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.290115801 +0000 UTC m=+145.937905133" watchObservedRunningTime="2025-09-29 19:11:44.322166921 +0000 UTC m=+145.969956253" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.322641 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-wn992" podStartSLOduration=7.322637565 podStartE2EDuration="7.322637565s" podCreationTimestamp="2025-09-29 19:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.320747855 +0000 UTC m=+145.968537187" watchObservedRunningTime="2025-09-29 19:11:44.322637565 +0000 UTC m=+145.970426897" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.372107 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" podStartSLOduration=126.372090594 podStartE2EDuration="2m6.372090594s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.365866994 +0000 UTC m=+146.013656326" watchObservedRunningTime="2025-09-29 19:11:44.372090594 +0000 UTC m=+146.019879926" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.391098 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-267cw" podStartSLOduration=126.391079083 podStartE2EDuration="2m6.391079083s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.386416834 +0000 UTC m=+146.034206176" watchObservedRunningTime="2025-09-29 19:11:44.391079083 +0000 UTC m=+146.038868425" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.416359 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.423079 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:44.923061731 +0000 UTC m=+146.570851153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.443949 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" podStartSLOduration=126.4439331 podStartE2EDuration="2m6.4439331s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.409839686 +0000 UTC m=+146.057629018" watchObservedRunningTime="2025-09-29 19:11:44.4439331 +0000 UTC m=+146.091722432" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.500897 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" podStartSLOduration=126.500878559 podStartE2EDuration="2m6.500878559s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.445905714 +0000 UTC m=+146.093695046" watchObservedRunningTime="2025-09-29 19:11:44.500878559 +0000 UTC m=+146.148667891" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.528849 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.529128 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.029113256 +0000 UTC m=+146.676902588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.581410 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" podStartSLOduration=126.581370744 podStartE2EDuration="2m6.581370744s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.539763228 +0000 UTC m=+146.187552560" watchObservedRunningTime="2025-09-29 19:11:44.581370744 +0000 UTC m=+146.229160076" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.583555 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" podStartSLOduration=126.583544893 podStartE2EDuration="2m6.583544893s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.579966669 +0000 UTC m=+146.227756001" watchObservedRunningTime="2025-09-29 19:11:44.583544893 +0000 UTC m=+146.231334225" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.630363 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.630779 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.1307676 +0000 UTC m=+146.778556932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.636538 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ww5bb" podStartSLOduration=126.636517914 podStartE2EDuration="2m6.636517914s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.624729837 +0000 UTC m=+146.272519169" watchObservedRunningTime="2025-09-29 19:11:44.636517914 +0000 UTC m=+146.284307246" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.666917 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6ksg" podStartSLOduration=126.666897971 podStartE2EDuration="2m6.666897971s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.661752715 +0000 UTC m=+146.309542047" watchObservedRunningTime="2025-09-29 19:11:44.666897971 +0000 UTC m=+146.314687303" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.702072 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qldrx" podStartSLOduration=126.702054959 podStartE2EDuration="2m6.702054959s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.699697894 +0000 UTC m=+146.347487236" watchObservedRunningTime="2025-09-29 19:11:44.702054959 +0000 UTC m=+146.349844291" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.731701 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.732059 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.232039052 +0000 UTC m=+146.879828384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.781740 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" podStartSLOduration=126.781716387 podStartE2EDuration="2m6.781716387s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.753874023 +0000 UTC m=+146.401663355" watchObservedRunningTime="2025-09-29 19:11:44.781716387 +0000 UTC m=+146.429505719" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.832682 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.833090 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.333072426 +0000 UTC m=+146.980861828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.849806 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" podStartSLOduration=126.849784343 podStartE2EDuration="2m6.849784343s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.848896484 +0000 UTC m=+146.496685816" watchObservedRunningTime="2025-09-29 19:11:44.849784343 +0000 UTC m=+146.497573685" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.852599 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" podStartSLOduration=126.852584533 podStartE2EDuration="2m6.852584533s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:44.820865075 +0000 UTC m=+146.468654407" watchObservedRunningTime="2025-09-29 19:11:44.852584533 +0000 UTC m=+146.500373865" Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.933843 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.933974 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.433951436 +0000 UTC m=+147.081740778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:44 crc kubenswrapper[4741]: I0929 19:11:44.934352 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:44 crc kubenswrapper[4741]: E0929 19:11:44.934794 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.434772992 +0000 UTC m=+147.082562324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.035851 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.035983 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.535948071 +0000 UTC m=+147.183737403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.036226 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.036527 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.536519469 +0000 UTC m=+147.184308791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.136892 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.137089 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.637056278 +0000 UTC m=+147.284845610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.137136 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.137460 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.63744683 +0000 UTC m=+147.285236162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.238256 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.238464 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.738436973 +0000 UTC m=+147.386226305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.238600 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.238862 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.738849786 +0000 UTC m=+147.386639118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.250717 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:45 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:45 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:45 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.250774 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.265978 4741 patch_prober.go:28] interesting pod/console-operator-58897d9998-4r5pc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.266036 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" podUID="5e2985ed-ad44-4d45-9b9b-ad1438579635" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.271981 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" event={"ID":"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6","Type":"ContainerStarted","Data":"b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.272625 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.274283 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-f99d7" event={"ID":"88c5801d-74cf-48c9-8443-8dcd8a743ff3","Type":"ContainerStarted","Data":"ec93837f6bdc144111695c12394067026e4e8e609f4b1157342ec7a0da46e485"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.277155 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hktj8" event={"ID":"a4235e5f-4b25-4e97-b13d-57fcde51d267","Type":"ContainerStarted","Data":"2505b34725fad10eb4856680b06119c549244950ebd5ede939f5c7aebf948e44"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.280049 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2tdpd" event={"ID":"965617c1-858f-4aa3-8c89-185f45434afd","Type":"ContainerStarted","Data":"9b5a026d34dcfaaf8827f3b192e1f1730b85f0abbeaf8be87fa3bf1875e137b7"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.281948 4741 generic.go:334] "Generic (PLEG): container finished" podID="b6394941-7bc5-45db-9404-6fa52d0fb10f" containerID="292180dc73a15c3eab4c0cf1929f8f62499d6f33cd171557c9522193af31e653" exitCode=0 Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.281981 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" event={"ID":"b6394941-7bc5-45db-9404-6fa52d0fb10f","Type":"ContainerDied","Data":"292180dc73a15c3eab4c0cf1929f8f62499d6f33cd171557c9522193af31e653"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.283954 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pcthr" event={"ID":"98c86725-7ab9-4054-9006-2e666f5b9576","Type":"ContainerStarted","Data":"d568ffcb5f3567306d0b2d608388efd9226ab2df2ae2d36486f1493253daa42b"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.285217 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" event={"ID":"9c4347d9-37c2-4c1e-806b-1c2927c1638b","Type":"ContainerStarted","Data":"135cfaf3b58cd7cb7762fe449f2a2a9e1d70512fb8b6250cc55222e2f8a28702"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.287263 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" event={"ID":"7030a932-f4cf-4ef3-9db5-9e3a7da7a547","Type":"ContainerStarted","Data":"eec164cd8ee30fff196f4cc1135793260d8af2b0823096cbc184d7f6f6a8b4fc"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.287299 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-25hcn" event={"ID":"7030a932-f4cf-4ef3-9db5-9e3a7da7a547","Type":"ContainerStarted","Data":"478e5d6554566d0440699544517c07ece09773eda4596c5320f6af85f27905f9"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.291549 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" event={"ID":"8036b4af-58a3-4e55-9643-4832021a71c5","Type":"ContainerStarted","Data":"8a5cc38847e35f0c51dfbcb8dd3d51f0481b340d3e377f6a7acee09ad116c5db"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.295525 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" event={"ID":"5e2aeccd-cc8d-47ff-801c-8ee9b243415a","Type":"ContainerStarted","Data":"3d38f093b58780e90dd6ec47e2e6bede4b4b2e3c2a82148c7137bdd96c2bfb50"} Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.296418 4741 patch_prober.go:28] interesting pod/downloads-7954f5f757-8xdlm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.296466 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8xdlm" podUID="bd769734-8dc9-445d-b482-5e8ef9357a0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.308978 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bdqsf" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.310351 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d674b" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.336626 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mllcn" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.339359 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" podStartSLOduration=127.339347403 podStartE2EDuration="2m7.339347403s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:45.33767366 +0000 UTC m=+146.985462992" watchObservedRunningTime="2025-09-29 19:11:45.339347403 +0000 UTC m=+146.987136735" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.340026 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.341841 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.841821213 +0000 UTC m=+147.489610545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.442595 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.443001 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:45.942985671 +0000 UTC m=+147.590775003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.544175 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.544364 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.044338086 +0000 UTC m=+147.692127418 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.544759 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.545081 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.04507344 +0000 UTC m=+147.692862772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.595488 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" podStartSLOduration=127.595465888 podStartE2EDuration="2m7.595465888s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:45.591717298 +0000 UTC m=+147.239506640" watchObservedRunningTime="2025-09-29 19:11:45.595465888 +0000 UTC m=+147.243255220" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.631514 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4r5pc" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.643657 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" podStartSLOduration=127.643638985 podStartE2EDuration="2m7.643638985s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:45.639839103 +0000 UTC m=+147.287628435" watchObservedRunningTime="2025-09-29 19:11:45.643638985 +0000 UTC m=+147.291428317" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.646254 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.646509 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.146484337 +0000 UTC m=+147.794273669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.646734 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.647001 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.146992812 +0000 UTC m=+147.794782144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.748193 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.748608 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.248587665 +0000 UTC m=+147.896376997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.849369 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.849475 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.849536 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.849896 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.349878598 +0000 UTC m=+147.997667930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.854705 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.867947 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.869945 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xjn62" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.950874 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.951149 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.951183 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:45 crc kubenswrapper[4741]: E0929 19:11:45.952363 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.452344008 +0000 UTC m=+148.100133350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.956524 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.959768 4741 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bzlbd container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Sep 29 19:11:45 crc kubenswrapper[4741]: [+]log ok Sep 29 19:11:45 crc kubenswrapper[4741]: [-]poststarthook/max-in-flight-filter failed: reason withheld Sep 29 19:11:45 crc kubenswrapper[4741]: [+]poststarthook/storage-object-count-tracker-hook ok Sep 29 19:11:45 crc kubenswrapper[4741]: [+]poststarthook/openshift.io-StartUserInformer ok Sep 29 19:11:45 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.959810 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" podUID="bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" containerName="oauth-openshift" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:45 crc kubenswrapper[4741]: I0929 19:11:45.960140 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.003652 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.020400 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.052582 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.052924 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.552911957 +0000 UTC m=+148.200701289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.096731 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.153755 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.153938 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.65391309 +0000 UTC m=+148.301702422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.154087 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.154414 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.654402086 +0000 UTC m=+148.302191418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.249378 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c26bp"] Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.256503 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.259868 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c26bp"] Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.260060 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:46 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:46 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:46 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.260096 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.260693 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.261892 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.261942 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.761917389 +0000 UTC m=+148.409706721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.324798 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" event={"ID":"9c4347d9-37c2-4c1e-806b-1c2927c1638b","Type":"ContainerStarted","Data":"0d73274a8c63d98764971e37136066230ce1b2e0dd5e2673c8f4dcaf59622476"} Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.324842 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" event={"ID":"9c4347d9-37c2-4c1e-806b-1c2927c1638b","Type":"ContainerStarted","Data":"f25b07c1ed7c2836a65d389dedc7483c156ef51b39059248303812ed667c6a7e"} Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.331062 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.332254 4741 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.366762 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-catalog-content\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.366894 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-utilities\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.367143 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lrb8\" (UniqueName: \"kubernetes.io/projected/c9bd836e-9017-45c5-9de1-d8bcba956088-kube-api-access-6lrb8\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.367254 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.367517 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.867505179 +0000 UTC m=+148.515294511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.440562 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kcsrb"] Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.441685 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.446854 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.460622 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kcsrb"] Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.472050 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.472250 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcbvz\" (UniqueName: \"kubernetes.io/projected/e92734a4-fea9-42b6-badf-dc94c29db4fc-kube-api-access-lcbvz\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.472308 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lrb8\" (UniqueName: \"kubernetes.io/projected/c9bd836e-9017-45c5-9de1-d8bcba956088-kube-api-access-6lrb8\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.472359 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-utilities\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.472405 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-catalog-content\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.472469 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-catalog-content\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.472507 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-utilities\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.472534 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:46.972509191 +0000 UTC m=+148.620298523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.473000 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-utilities\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.473152 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-catalog-content\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.497194 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lrb8\" (UniqueName: \"kubernetes.io/projected/c9bd836e-9017-45c5-9de1-d8bcba956088-kube-api-access-6lrb8\") pod \"community-operators-c26bp\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.573405 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcbvz\" (UniqueName: \"kubernetes.io/projected/e92734a4-fea9-42b6-badf-dc94c29db4fc-kube-api-access-lcbvz\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.573489 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.573525 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-utilities\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.573560 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-catalog-content\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.574034 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-catalog-content\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.574614 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.074602639 +0000 UTC m=+148.722391971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.575137 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-utilities\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.593078 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.597050 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcbvz\" (UniqueName: \"kubernetes.io/projected/e92734a4-fea9-42b6-badf-dc94c29db4fc-kube-api-access-lcbvz\") pod \"certified-operators-kcsrb\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.649369 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lfcss"] Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.666594 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfcss"] Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.666845 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.675022 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.675540 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.175499789 +0000 UTC m=+148.823289131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.733734 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.776499 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6394941-7bc5-45db-9404-6fa52d0fb10f-secret-volume\") pod \"b6394941-7bc5-45db-9404-6fa52d0fb10f\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.776553 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6394941-7bc5-45db-9404-6fa52d0fb10f-config-volume\") pod \"b6394941-7bc5-45db-9404-6fa52d0fb10f\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.776576 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgmsj\" (UniqueName: \"kubernetes.io/projected/b6394941-7bc5-45db-9404-6fa52d0fb10f-kube-api-access-vgmsj\") pod \"b6394941-7bc5-45db-9404-6fa52d0fb10f\" (UID: \"b6394941-7bc5-45db-9404-6fa52d0fb10f\") " Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.776738 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-utilities\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.776763 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.776782 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-catalog-content\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.776805 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwm6v\" (UniqueName: \"kubernetes.io/projected/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-kube-api-access-xwm6v\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.781826 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6394941-7bc5-45db-9404-6fa52d0fb10f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b6394941-7bc5-45db-9404-6fa52d0fb10f" (UID: "b6394941-7bc5-45db-9404-6fa52d0fb10f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.782294 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6394941-7bc5-45db-9404-6fa52d0fb10f-config-volume" (OuterVolumeSpecName: "config-volume") pod "b6394941-7bc5-45db-9404-6fa52d0fb10f" (UID: "b6394941-7bc5-45db-9404-6fa52d0fb10f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.782716 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.783266 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.28324825 +0000 UTC m=+148.931037582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.797320 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6394941-7bc5-45db-9404-6fa52d0fb10f-kube-api-access-vgmsj" (OuterVolumeSpecName: "kube-api-access-vgmsj") pod "b6394941-7bc5-45db-9404-6fa52d0fb10f" (UID: "b6394941-7bc5-45db-9404-6fa52d0fb10f"). InnerVolumeSpecName "kube-api-access-vgmsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.833076 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6ggfh"] Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.833463 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6394941-7bc5-45db-9404-6fa52d0fb10f" containerName="collect-profiles" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.833545 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6394941-7bc5-45db-9404-6fa52d0fb10f" containerName="collect-profiles" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.833713 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6394941-7bc5-45db-9404-6fa52d0fb10f" containerName="collect-profiles" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.834433 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.859595 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ggfh"] Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.877780 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.878466 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-utilities\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.878515 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.378485078 +0000 UTC m=+149.026274410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.878662 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d6d6\" (UniqueName: \"kubernetes.io/projected/3a5e870e-8a78-4835-b453-764cae6e4b41-kube-api-access-4d6d6\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.878753 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-utilities\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.878841 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.878915 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-catalog-content\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.878986 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-catalog-content\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.879069 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwm6v\" (UniqueName: \"kubernetes.io/projected/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-kube-api-access-xwm6v\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.879154 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6394941-7bc5-45db-9404-6fa52d0fb10f-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.879213 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6394941-7bc5-45db-9404-6fa52d0fb10f-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.879273 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgmsj\" (UniqueName: \"kubernetes.io/projected/b6394941-7bc5-45db-9404-6fa52d0fb10f-kube-api-access-vgmsj\") on node \"crc\" DevicePath \"\"" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.879969 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-utilities\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.880886 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-catalog-content\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.890865 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.380274765 +0000 UTC m=+149.028064097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.917269 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwm6v\" (UniqueName: \"kubernetes.io/projected/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-kube-api-access-xwm6v\") pod \"community-operators-lfcss\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.982086 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.982242 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.482218848 +0000 UTC m=+149.130008180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.982464 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.982517 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-catalog-content\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.982577 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-utilities\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.982655 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d6d6\" (UniqueName: \"kubernetes.io/projected/3a5e870e-8a78-4835-b453-764cae6e4b41-kube-api-access-4d6d6\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: E0929 19:11:46.982772 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.482763206 +0000 UTC m=+149.130552538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.983122 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-catalog-content\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:46 crc kubenswrapper[4741]: I0929 19:11:46.985536 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-utilities\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.010182 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d6d6\" (UniqueName: \"kubernetes.io/projected/3a5e870e-8a78-4835-b453-764cae6e4b41-kube-api-access-4d6d6\") pod \"certified-operators-6ggfh\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.056893 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.072756 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c26bp"] Sep 29 19:11:47 crc kubenswrapper[4741]: W0929 19:11:47.083753 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9bd836e_9017_45c5_9de1_d8bcba956088.slice/crio-3ffe02a03ee97a11c66e710e5212465ad85046e55013133fa27f4b2e1151edc5 WatchSource:0}: Error finding container 3ffe02a03ee97a11c66e710e5212465ad85046e55013133fa27f4b2e1151edc5: Status 404 returned error can't find the container with id 3ffe02a03ee97a11c66e710e5212465ad85046e55013133fa27f4b2e1151edc5 Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.083801 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:47 crc kubenswrapper[4741]: E0929 19:11:47.083879 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.583859982 +0000 UTC m=+149.231649314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.084050 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:47 crc kubenswrapper[4741]: E0929 19:11:47.084292 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.584285276 +0000 UTC m=+149.232074608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.110514 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kcsrb"] Sep 29 19:11:47 crc kubenswrapper[4741]: W0929 19:11:47.125813 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode92734a4_fea9_42b6_badf_dc94c29db4fc.slice/crio-a82ef4d99c1fbd831d511122e1d38a67e148c51579f908ea888ab595f7d159ee WatchSource:0}: Error finding container a82ef4d99c1fbd831d511122e1d38a67e148c51579f908ea888ab595f7d159ee: Status 404 returned error can't find the container with id a82ef4d99c1fbd831d511122e1d38a67e148c51579f908ea888ab595f7d159ee Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.164037 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.184896 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:47 crc kubenswrapper[4741]: E0929 19:11:47.185688 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.685671302 +0000 UTC m=+149.333460634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.260606 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:47 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:47 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:47 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.260660 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.291527 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:47 crc kubenswrapper[4741]: E0929 19:11:47.291876 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 19:11:47.791861702 +0000 UTC m=+149.439651034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-d4ltl" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.318595 4741 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-09-29T19:11:46.333608331Z","Handler":null,"Name":""} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.323338 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfcss"] Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.324566 4741 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.324597 4741 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.337834 4741 generic.go:334] "Generic (PLEG): container finished" podID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerID="959f96147a6cb8490a090f8eae463193ac9fcc7203f4b8792eefa393fcff8f5f" exitCode=0 Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.337952 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c26bp" event={"ID":"c9bd836e-9017-45c5-9de1-d8bcba956088","Type":"ContainerDied","Data":"959f96147a6cb8490a090f8eae463193ac9fcc7203f4b8792eefa393fcff8f5f"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.338033 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c26bp" event={"ID":"c9bd836e-9017-45c5-9de1-d8bcba956088","Type":"ContainerStarted","Data":"3ffe02a03ee97a11c66e710e5212465ad85046e55013133fa27f4b2e1151edc5"} Sep 29 19:11:47 crc kubenswrapper[4741]: W0929 19:11:47.342803 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84b44513_b985_42e9_ba2c_1c3af7c6ebd1.slice/crio-a3abb031b69717f8e1ea5a7c1e49af245e7140f48bbb6cc08ad232671038d451 WatchSource:0}: Error finding container a3abb031b69717f8e1ea5a7c1e49af245e7140f48bbb6cc08ad232671038d451: Status 404 returned error can't find the container with id a3abb031b69717f8e1ea5a7c1e49af245e7140f48bbb6cc08ad232671038d451 Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.344881 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.345191 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1749df7f19de4fb3ff28051fd927633a849db9ba342f95acf107ad00fc74df87"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.345226 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d499728cfc3ed0de3acdcee373545e0c39f78fd544411e35fcb2e3a0c3db8c59"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.347939 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4420ef9780ed1887467799b8491d174258322f6afba69ee4634dd47c57bc1952"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.347970 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a4f715aa0361d56ce2e0d8b3a2fa6bd963af444e27cce4e2a46a47e1c58f0c8f"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.348447 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.350690 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" event={"ID":"9c4347d9-37c2-4c1e-806b-1c2927c1638b","Type":"ContainerStarted","Data":"6559a60ffb248e5f8cf7fe094bee3447e6969e9f59093f90969ba4453278c4a9"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.352843 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"df50ecafb8353540eedc57f1d01641a09e0ca1dbfdb74248e256d5f3fd11d2da"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.352882 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2ac6d97ed44e89bc82cb124e67b2ce79c339acb94d1c9abeb5ee4291fb807650"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.363100 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kcsrb" event={"ID":"e92734a4-fea9-42b6-badf-dc94c29db4fc","Type":"ContainerStarted","Data":"110f431df0cca0b8961bdd50adf9023e2516a4935afe31b110b154779bf366a5"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.363129 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kcsrb" event={"ID":"e92734a4-fea9-42b6-badf-dc94c29db4fc","Type":"ContainerStarted","Data":"a82ef4d99c1fbd831d511122e1d38a67e148c51579f908ea888ab595f7d159ee"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.365994 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.366951 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk" event={"ID":"b6394941-7bc5-45db-9404-6fa52d0fb10f","Type":"ContainerDied","Data":"51d0652cb7115bd7d2bc2ac922a38bccac18abe2e5e26efa2530783393291c82"} Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.366977 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d0652cb7115bd7d2bc2ac922a38bccac18abe2e5e26efa2530783393291c82" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.382759 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-c6qb4" podStartSLOduration=10.38274069 podStartE2EDuration="10.38274069s" podCreationTimestamp="2025-09-29 19:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:47.379403963 +0000 UTC m=+149.027193305" watchObservedRunningTime="2025-09-29 19:11:47.38274069 +0000 UTC m=+149.030530022" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.394665 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.405305 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.445820 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ggfh"] Sep 29 19:11:47 crc kubenswrapper[4741]: W0929 19:11:47.489255 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a5e870e_8a78_4835_b453_764cae6e4b41.slice/crio-f0e11efb117f5d56943ee5472c4a16b5bce26e7be2c5736e09f573bb22a10615 WatchSource:0}: Error finding container f0e11efb117f5d56943ee5472c4a16b5bce26e7be2c5736e09f573bb22a10615: Status 404 returned error can't find the container with id f0e11efb117f5d56943ee5472c4a16b5bce26e7be2c5736e09f573bb22a10615 Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.503253 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.511576 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.511625 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.555896 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-d4ltl\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.559441 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.560236 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.572407 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.572599 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.575922 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.621888 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.705813 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c11991f2-4d1f-4c56-81aa-09bbe9507683-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.705897 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11991f2-4d1f-4c56-81aa-09bbe9507683-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.802913 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-d4ltl"] Sep 29 19:11:47 crc kubenswrapper[4741]: W0929 19:11:47.810120 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf05d1ab9_cf4d_4a05_a4bd_2d029564c266.slice/crio-9780b9423bf6cdd730f796fdcc223cf3d4fa7b87fb60b316a7624fbd36c05e63 WatchSource:0}: Error finding container 9780b9423bf6cdd730f796fdcc223cf3d4fa7b87fb60b316a7624fbd36c05e63: Status 404 returned error can't find the container with id 9780b9423bf6cdd730f796fdcc223cf3d4fa7b87fb60b316a7624fbd36c05e63 Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.815743 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11991f2-4d1f-4c56-81aa-09bbe9507683-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.815841 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c11991f2-4d1f-4c56-81aa-09bbe9507683-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.815897 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c11991f2-4d1f-4c56-81aa-09bbe9507683-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.833666 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11991f2-4d1f-4c56-81aa-09bbe9507683-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:47 crc kubenswrapper[4741]: I0929 19:11:47.930155 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.128662 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.249883 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:48 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:48 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:48 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.249940 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.382665 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" event={"ID":"f05d1ab9-cf4d-4a05-a4bd-2d029564c266","Type":"ContainerStarted","Data":"6c4baea219594446fc8f4e051043f3678f8e20a2a368cd2cbbeaf711c2804e6f"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.382713 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" event={"ID":"f05d1ab9-cf4d-4a05-a4bd-2d029564c266","Type":"ContainerStarted","Data":"9780b9423bf6cdd730f796fdcc223cf3d4fa7b87fb60b316a7624fbd36c05e63"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.383801 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.388245 4741 generic.go:334] "Generic (PLEG): container finished" podID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerID="110f431df0cca0b8961bdd50adf9023e2516a4935afe31b110b154779bf366a5" exitCode=0 Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.388308 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kcsrb" event={"ID":"e92734a4-fea9-42b6-badf-dc94c29db4fc","Type":"ContainerDied","Data":"110f431df0cca0b8961bdd50adf9023e2516a4935afe31b110b154779bf366a5"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.390266 4741 generic.go:334] "Generic (PLEG): container finished" podID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerID="d9054f9e7f8ccd4e483bfbb507f845674d9b28f02553e9eb743bbcb5ae39a42f" exitCode=0 Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.390306 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfcss" event={"ID":"84b44513-b985-42e9-ba2c-1c3af7c6ebd1","Type":"ContainerDied","Data":"d9054f9e7f8ccd4e483bfbb507f845674d9b28f02553e9eb743bbcb5ae39a42f"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.390324 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfcss" event={"ID":"84b44513-b985-42e9-ba2c-1c3af7c6ebd1","Type":"ContainerStarted","Data":"a3abb031b69717f8e1ea5a7c1e49af245e7140f48bbb6cc08ad232671038d451"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.392356 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c11991f2-4d1f-4c56-81aa-09bbe9507683","Type":"ContainerStarted","Data":"26b2152ab6c08e5e86dec784a9882e45a50f3fa3ddf0d33fbfbf828f27018649"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.397604 4741 generic.go:334] "Generic (PLEG): container finished" podID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerID="b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27" exitCode=0 Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.397671 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ggfh" event={"ID":"3a5e870e-8a78-4835-b453-764cae6e4b41","Type":"ContainerDied","Data":"b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.397713 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ggfh" event={"ID":"3a5e870e-8a78-4835-b453-764cae6e4b41","Type":"ContainerStarted","Data":"f0e11efb117f5d56943ee5472c4a16b5bce26e7be2c5736e09f573bb22a10615"} Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.410731 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" podStartSLOduration=130.410702139 podStartE2EDuration="2m10.410702139s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:11:48.40295085 +0000 UTC m=+150.050740182" watchObservedRunningTime="2025-09-29 19:11:48.410702139 +0000 UTC m=+150.058491471" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.440514 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2vmvz"] Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.441968 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.442002 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vmvz"] Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.443815 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.635003 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndfnk\" (UniqueName: \"kubernetes.io/projected/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-kube-api-access-ndfnk\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.635377 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-catalog-content\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.635477 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-utilities\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.736177 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndfnk\" (UniqueName: \"kubernetes.io/projected/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-kube-api-access-ndfnk\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.736329 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-catalog-content\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.736404 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-utilities\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.736977 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-utilities\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.737023 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-catalog-content\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.755008 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndfnk\" (UniqueName: \"kubernetes.io/projected/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-kube-api-access-ndfnk\") pod \"redhat-marketplace-2vmvz\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.770557 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.829360 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bf8jv"] Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.830748 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.851179 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf8jv"] Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.939310 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-utilities\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.939348 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjnwm\" (UniqueName: \"kubernetes.io/projected/b13e515d-1615-4338-b8b3-1b08e4d3faf4-kube-api-access-jjnwm\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:48 crc kubenswrapper[4741]: I0929 19:11:48.939502 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-catalog-content\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.041308 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-utilities\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.041774 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjnwm\" (UniqueName: \"kubernetes.io/projected/b13e515d-1615-4338-b8b3-1b08e4d3faf4-kube-api-access-jjnwm\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.041893 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-catalog-content\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.042122 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-utilities\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.042341 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-catalog-content\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.073243 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vmvz"] Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.073358 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjnwm\" (UniqueName: \"kubernetes.io/projected/b13e515d-1615-4338-b8b3-1b08e4d3faf4-kube-api-access-jjnwm\") pod \"redhat-marketplace-bf8jv\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: W0929 19:11:49.091643 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f69b21c_ce1f_4c27_8dcf_f71d2afcfee9.slice/crio-779fb826d31d42bf2f2367020715fc7b009418ddce3be36e3f5955a410319add WatchSource:0}: Error finding container 779fb826d31d42bf2f2367020715fc7b009418ddce3be36e3f5955a410319add: Status 404 returned error can't find the container with id 779fb826d31d42bf2f2367020715fc7b009418ddce3be36e3f5955a410319add Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.106034 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.161208 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.252855 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:49 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:49 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:49 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.253516 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.395868 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf8jv"] Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.404708 4741 generic.go:334] "Generic (PLEG): container finished" podID="c11991f2-4d1f-4c56-81aa-09bbe9507683" containerID="5d0071edd5865f3125307801ae8f23f9277346f9f791f529eaaeed8a7e4e5865" exitCode=0 Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.404767 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c11991f2-4d1f-4c56-81aa-09bbe9507683","Type":"ContainerDied","Data":"5d0071edd5865f3125307801ae8f23f9277346f9f791f529eaaeed8a7e4e5865"} Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.406475 4741 generic.go:334] "Generic (PLEG): container finished" podID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerID="3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f" exitCode=0 Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.406721 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vmvz" event={"ID":"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9","Type":"ContainerDied","Data":"3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f"} Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.406770 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vmvz" event={"ID":"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9","Type":"ContainerStarted","Data":"779fb826d31d42bf2f2367020715fc7b009418ddce3be36e3f5955a410319add"} Sep 29 19:11:49 crc kubenswrapper[4741]: W0929 19:11:49.412505 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb13e515d_1615_4338_b8b3_1b08e4d3faf4.slice/crio-5fb45479d4f59bf84f3d11e55e871febc2d8623331bdefab496440a069d4d989 WatchSource:0}: Error finding container 5fb45479d4f59bf84f3d11e55e871febc2d8623331bdefab496440a069d4d989: Status 404 returned error can't find the container with id 5fb45479d4f59bf84f3d11e55e871febc2d8623331bdefab496440a069d4d989 Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.429541 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rdvm7"] Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.430716 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.434338 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.443852 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdvm7"] Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.557680 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-utilities\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.557989 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-catalog-content\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.558080 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gtk9\" (UniqueName: \"kubernetes.io/projected/bbbbb850-78fa-47f0-8339-24148d9c610e-kube-api-access-9gtk9\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.659124 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gtk9\" (UniqueName: \"kubernetes.io/projected/bbbbb850-78fa-47f0-8339-24148d9c610e-kube-api-access-9gtk9\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.659251 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-utilities\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.659341 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-catalog-content\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.659920 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-utilities\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.659952 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-catalog-content\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.680268 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gtk9\" (UniqueName: \"kubernetes.io/projected/bbbbb850-78fa-47f0-8339-24148d9c610e-kube-api-access-9gtk9\") pod \"redhat-operators-rdvm7\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.753885 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.827380 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jdnwf"] Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.828708 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.838742 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdnwf"] Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.851961 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.851999 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.858780 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.858833 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.858952 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.870115 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.884026 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-utilities\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.884362 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-catalog-content\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.884477 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l24sk\" (UniqueName: \"kubernetes.io/projected/8e4390f3-92f6-4639-81ad-c1d11073309e-kube-api-access-l24sk\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.929371 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.929422 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.931856 4741 patch_prober.go:28] interesting pod/console-f9d7485db-jd7m4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.931923 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-jd7m4" podUID="932a4f81-a3d7-480f-a3bb-648b2e3bbd31" containerName="console" probeResult="failure" output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.985926 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-catalog-content\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.985993 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l24sk\" (UniqueName: \"kubernetes.io/projected/8e4390f3-92f6-4639-81ad-c1d11073309e-kube-api-access-l24sk\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.986130 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-utilities\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.986887 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-utilities\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:49 crc kubenswrapper[4741]: I0929 19:11:49.987702 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-catalog-content\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.004174 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l24sk\" (UniqueName: \"kubernetes.io/projected/8e4390f3-92f6-4639-81ad-c1d11073309e-kube-api-access-l24sk\") pod \"redhat-operators-jdnwf\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.040944 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdvm7"] Sep 29 19:11:50 crc kubenswrapper[4741]: W0929 19:11:50.073834 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbbbb850_78fa_47f0_8339_24148d9c610e.slice/crio-15c7a9c498f0361f92aec09090766c78cf86c2e0c0e89e7065004c9d18575019 WatchSource:0}: Error finding container 15c7a9c498f0361f92aec09090766c78cf86c2e0c0e89e7065004c9d18575019: Status 404 returned error can't find the container with id 15c7a9c498f0361f92aec09090766c78cf86c2e0c0e89e7065004c9d18575019 Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.145479 4741 patch_prober.go:28] interesting pod/downloads-7954f5f757-8xdlm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.145808 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8xdlm" podUID="bd769734-8dc9-445d-b482-5e8ef9357a0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.145568 4741 patch_prober.go:28] interesting pod/downloads-7954f5f757-8xdlm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.146217 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8xdlm" podUID="bd769734-8dc9-445d-b482-5e8ef9357a0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.155781 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.247063 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.252002 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:50 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:50 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:50 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.252045 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.417161 4741 generic.go:334] "Generic (PLEG): container finished" podID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerID="15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946" exitCode=0 Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.417227 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf8jv" event={"ID":"b13e515d-1615-4338-b8b3-1b08e4d3faf4","Type":"ContainerDied","Data":"15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946"} Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.417252 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf8jv" event={"ID":"b13e515d-1615-4338-b8b3-1b08e4d3faf4","Type":"ContainerStarted","Data":"5fb45479d4f59bf84f3d11e55e871febc2d8623331bdefab496440a069d4d989"} Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.419638 4741 generic.go:334] "Generic (PLEG): container finished" podID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerID="846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376" exitCode=0 Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.419676 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdvm7" event={"ID":"bbbbb850-78fa-47f0-8339-24148d9c610e","Type":"ContainerDied","Data":"846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376"} Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.419717 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdvm7" event={"ID":"bbbbb850-78fa-47f0-8339-24148d9c610e","Type":"ContainerStarted","Data":"15c7a9c498f0361f92aec09090766c78cf86c2e0c0e89e7065004c9d18575019"} Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.425553 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-h8d49" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.426301 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dz9j9" Sep 29 19:11:50 crc kubenswrapper[4741]: I0929 19:11:50.584697 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdnwf"] Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.007951 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.118378 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c11991f2-4d1f-4c56-81aa-09bbe9507683-kubelet-dir\") pod \"c11991f2-4d1f-4c56-81aa-09bbe9507683\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.118483 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11991f2-4d1f-4c56-81aa-09bbe9507683-kube-api-access\") pod \"c11991f2-4d1f-4c56-81aa-09bbe9507683\" (UID: \"c11991f2-4d1f-4c56-81aa-09bbe9507683\") " Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.119025 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c11991f2-4d1f-4c56-81aa-09bbe9507683-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c11991f2-4d1f-4c56-81aa-09bbe9507683" (UID: "c11991f2-4d1f-4c56-81aa-09bbe9507683"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.142896 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c11991f2-4d1f-4c56-81aa-09bbe9507683-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c11991f2-4d1f-4c56-81aa-09bbe9507683" (UID: "c11991f2-4d1f-4c56-81aa-09bbe9507683"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.219504 4741 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c11991f2-4d1f-4c56-81aa-09bbe9507683-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.219831 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11991f2-4d1f-4c56-81aa-09bbe9507683-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.251933 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:51 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:51 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:51 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.252008 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.377741 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 19:11:51 crc kubenswrapper[4741]: E0929 19:11:51.377959 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11991f2-4d1f-4c56-81aa-09bbe9507683" containerName="pruner" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.377970 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11991f2-4d1f-4c56-81aa-09bbe9507683" containerName="pruner" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.378065 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c11991f2-4d1f-4c56-81aa-09bbe9507683" containerName="pruner" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.378437 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.393278 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.393516 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.399348 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.452660 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c11991f2-4d1f-4c56-81aa-09bbe9507683","Type":"ContainerDied","Data":"26b2152ab6c08e5e86dec784a9882e45a50f3fa3ddf0d33fbfbf828f27018649"} Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.452704 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26b2152ab6c08e5e86dec784a9882e45a50f3fa3ddf0d33fbfbf828f27018649" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.452764 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.467317 4741 generic.go:334] "Generic (PLEG): container finished" podID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerID="01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d" exitCode=0 Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.468075 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdnwf" event={"ID":"8e4390f3-92f6-4639-81ad-c1d11073309e","Type":"ContainerDied","Data":"01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d"} Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.468109 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdnwf" event={"ID":"8e4390f3-92f6-4639-81ad-c1d11073309e","Type":"ContainerStarted","Data":"70dbe3e201770588765f81086f80ca7efdaea2fe37bd570afa1bcf3d45405f39"} Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.524667 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e827b-32c0-4af9-8f81-34a549a3c871-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.524731 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e827b-32c0-4af9-8f81-34a549a3c871-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.626622 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e827b-32c0-4af9-8f81-34a549a3c871-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.626667 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e827b-32c0-4af9-8f81-34a549a3c871-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.628063 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e827b-32c0-4af9-8f81-34a549a3c871-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.648967 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e827b-32c0-4af9-8f81-34a549a3c871-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:51 crc kubenswrapper[4741]: I0929 19:11:51.705707 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:11:52 crc kubenswrapper[4741]: I0929 19:11:52.144268 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 19:11:52 crc kubenswrapper[4741]: I0929 19:11:52.250842 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:52 crc kubenswrapper[4741]: [-]has-synced failed: reason withheld Sep 29 19:11:52 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:52 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:52 crc kubenswrapper[4741]: I0929 19:11:52.250947 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:52 crc kubenswrapper[4741]: I0929 19:11:52.314812 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-f99d7" Sep 29 19:11:52 crc kubenswrapper[4741]: I0929 19:11:52.476808 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e7e827b-32c0-4af9-8f81-34a549a3c871","Type":"ContainerStarted","Data":"9b1a7c78c7a543b28b4c34ca87af86e28992964ab3aafc4d93fa297ba0a36e31"} Sep 29 19:11:53 crc kubenswrapper[4741]: I0929 19:11:53.250179 4741 patch_prober.go:28] interesting pod/router-default-5444994796-ww5bb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 19:11:53 crc kubenswrapper[4741]: [+]has-synced ok Sep 29 19:11:53 crc kubenswrapper[4741]: [+]process-running ok Sep 29 19:11:53 crc kubenswrapper[4741]: healthz check failed Sep 29 19:11:53 crc kubenswrapper[4741]: I0929 19:11:53.250480 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ww5bb" podUID="7b4710bf-98cc-4ac1-9356-da1249819e88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 19:11:53 crc kubenswrapper[4741]: I0929 19:11:53.503507 4741 generic.go:334] "Generic (PLEG): container finished" podID="9e7e827b-32c0-4af9-8f81-34a549a3c871" containerID="e4ef2ef78258501135fe0921ab6cbc082f9dccef6d71c2b569df08feaee46832" exitCode=0 Sep 29 19:11:53 crc kubenswrapper[4741]: I0929 19:11:53.503563 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e7e827b-32c0-4af9-8f81-34a549a3c871","Type":"ContainerDied","Data":"e4ef2ef78258501135fe0921ab6cbc082f9dccef6d71c2b569df08feaee46832"} Sep 29 19:11:54 crc kubenswrapper[4741]: I0929 19:11:54.250902 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:54 crc kubenswrapper[4741]: I0929 19:11:54.252783 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ww5bb" Sep 29 19:11:59 crc kubenswrapper[4741]: I0929 19:11:59.933660 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:11:59 crc kubenswrapper[4741]: I0929 19:11:59.938101 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:12:00 crc kubenswrapper[4741]: I0929 19:12:00.164820 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8xdlm" Sep 29 19:12:00 crc kubenswrapper[4741]: I0929 19:12:00.169653 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:12:00 crc kubenswrapper[4741]: I0929 19:12:00.176158 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58436159-f44b-45ad-98f9-a60bc389bc7b-metrics-certs\") pod \"network-metrics-daemon-7krvl\" (UID: \"58436159-f44b-45ad-98f9-a60bc389bc7b\") " pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:12:00 crc kubenswrapper[4741]: I0929 19:12:00.412509 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7krvl" Sep 29 19:12:01 crc kubenswrapper[4741]: I0929 19:12:01.739655 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:12:01 crc kubenswrapper[4741]: I0929 19:12:01.740037 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.437500 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.565429 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e7e827b-32c0-4af9-8f81-34a549a3c871","Type":"ContainerDied","Data":"9b1a7c78c7a543b28b4c34ca87af86e28992964ab3aafc4d93fa297ba0a36e31"} Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.565475 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b1a7c78c7a543b28b4c34ca87af86e28992964ab3aafc4d93fa297ba0a36e31" Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.565540 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.608876 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e827b-32c0-4af9-8f81-34a549a3c871-kube-api-access\") pod \"9e7e827b-32c0-4af9-8f81-34a549a3c871\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.608993 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e827b-32c0-4af9-8f81-34a549a3c871-kubelet-dir\") pod \"9e7e827b-32c0-4af9-8f81-34a549a3c871\" (UID: \"9e7e827b-32c0-4af9-8f81-34a549a3c871\") " Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.609153 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e7e827b-32c0-4af9-8f81-34a549a3c871-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e7e827b-32c0-4af9-8f81-34a549a3c871" (UID: "9e7e827b-32c0-4af9-8f81-34a549a3c871"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.609356 4741 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e827b-32c0-4af9-8f81-34a549a3c871-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.613881 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e7e827b-32c0-4af9-8f81-34a549a3c871-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e7e827b-32c0-4af9-8f81-34a549a3c871" (UID: "9e7e827b-32c0-4af9-8f81-34a549a3c871"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:12:03 crc kubenswrapper[4741]: I0929 19:12:03.710668 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e827b-32c0-4af9-8f81-34a549a3c871-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:07 crc kubenswrapper[4741]: I0929 19:12:07.626663 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:12:16 crc kubenswrapper[4741]: I0929 19:12:16.246346 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 19:12:18 crc kubenswrapper[4741]: E0929 19:12:18.947180 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 29 19:12:18 crc kubenswrapper[4741]: E0929 19:12:18.947373 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ndfnk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2vmvz_openshift-marketplace(8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:12:18 crc kubenswrapper[4741]: E0929 19:12:18.948610 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2vmvz" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" Sep 29 19:12:19 crc kubenswrapper[4741]: E0929 19:12:19.993703 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 29 19:12:19 crc kubenswrapper[4741]: E0929 19:12:19.994702 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gtk9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rdvm7_openshift-marketplace(bbbbb850-78fa-47f0-8339-24148d9c610e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:12:19 crc kubenswrapper[4741]: E0929 19:12:19.996081 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rdvm7" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" Sep 29 19:12:20 crc kubenswrapper[4741]: I0929 19:12:20.285736 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-866mj" Sep 29 19:12:20 crc kubenswrapper[4741]: E0929 19:12:20.826630 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 29 19:12:20 crc kubenswrapper[4741]: E0929 19:12:20.826782 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6lrb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-c26bp_openshift-marketplace(c9bd836e-9017-45c5-9de1-d8bcba956088): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:12:20 crc kubenswrapper[4741]: E0929 19:12:20.828098 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-c26bp" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.086915 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-c26bp" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.086919 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rdvm7" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.087087 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2vmvz" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.169010 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.169724 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jjnwm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bf8jv_openshift-marketplace(b13e515d-1615-4338-b8b3-1b08e4d3faf4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.170985 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bf8jv" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.178518 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.178718 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l24sk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jdnwf_openshift-marketplace(8e4390f3-92f6-4639-81ad-c1d11073309e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:12:22 crc kubenswrapper[4741]: E0929 19:12:22.179881 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jdnwf" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.246619 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jdnwf" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.247175 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bf8jv" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.319473 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.319613 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4d6d6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6ggfh_openshift-marketplace(3a5e870e-8a78-4835-b453-764cae6e4b41): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.322473 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6ggfh" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.339484 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.339634 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lcbvz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kcsrb_openshift-marketplace(e92734a4-fea9-42b6-badf-dc94c29db4fc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.340780 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kcsrb" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" Sep 29 19:12:23 crc kubenswrapper[4741]: I0929 19:12:23.651848 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7krvl"] Sep 29 19:12:23 crc kubenswrapper[4741]: I0929 19:12:23.664064 4741 generic.go:334] "Generic (PLEG): container finished" podID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerID="443a784b4511ac91e0a98c6e73658191492ae729f7a1211245c601544e355022" exitCode=0 Sep 29 19:12:23 crc kubenswrapper[4741]: I0929 19:12:23.665408 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfcss" event={"ID":"84b44513-b985-42e9-ba2c-1c3af7c6ebd1","Type":"ContainerDied","Data":"443a784b4511ac91e0a98c6e73658191492ae729f7a1211245c601544e355022"} Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.665527 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6ggfh" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" Sep 29 19:12:23 crc kubenswrapper[4741]: E0929 19:12:23.667712 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kcsrb" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" Sep 29 19:12:23 crc kubenswrapper[4741]: W0929 19:12:23.711072 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58436159_f44b_45ad_98f9_a60bc389bc7b.slice/crio-a6ca2f6aba25df4020f2894a9fa18fbd783f5746ac0cabaa59820646e90c4f64 WatchSource:0}: Error finding container a6ca2f6aba25df4020f2894a9fa18fbd783f5746ac0cabaa59820646e90c4f64: Status 404 returned error can't find the container with id a6ca2f6aba25df4020f2894a9fa18fbd783f5746ac0cabaa59820646e90c4f64 Sep 29 19:12:24 crc kubenswrapper[4741]: I0929 19:12:24.672256 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7krvl" event={"ID":"58436159-f44b-45ad-98f9-a60bc389bc7b","Type":"ContainerStarted","Data":"db7bf0647e273dd37f1820ab1a6045612a13ebceccfb6b3ad5474d657c21ec1c"} Sep 29 19:12:24 crc kubenswrapper[4741]: I0929 19:12:24.672782 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7krvl" event={"ID":"58436159-f44b-45ad-98f9-a60bc389bc7b","Type":"ContainerStarted","Data":"722bcce2935d2b900d232943bc54a94e5e41f8a334d2c84440891e8a79510642"} Sep 29 19:12:24 crc kubenswrapper[4741]: I0929 19:12:24.672798 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7krvl" event={"ID":"58436159-f44b-45ad-98f9-a60bc389bc7b","Type":"ContainerStarted","Data":"a6ca2f6aba25df4020f2894a9fa18fbd783f5746ac0cabaa59820646e90c4f64"} Sep 29 19:12:24 crc kubenswrapper[4741]: I0929 19:12:24.675349 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfcss" event={"ID":"84b44513-b985-42e9-ba2c-1c3af7c6ebd1","Type":"ContainerStarted","Data":"d06c4023687ee93b59c35ba59f85be2f0ea5cc762e5d0b64e4112553f2dfbe99"} Sep 29 19:12:24 crc kubenswrapper[4741]: I0929 19:12:24.716810 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7krvl" podStartSLOduration=166.716791831 podStartE2EDuration="2m46.716791831s" podCreationTimestamp="2025-09-29 19:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:12:24.697812072 +0000 UTC m=+186.345601444" watchObservedRunningTime="2025-09-29 19:12:24.716791831 +0000 UTC m=+186.364581163" Sep 29 19:12:24 crc kubenswrapper[4741]: I0929 19:12:24.717765 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lfcss" podStartSLOduration=2.954221161 podStartE2EDuration="38.717760061s" podCreationTimestamp="2025-09-29 19:11:46 +0000 UTC" firstStartedPulling="2025-09-29 19:11:48.391685208 +0000 UTC m=+150.039474540" lastFinishedPulling="2025-09-29 19:12:24.155224108 +0000 UTC m=+185.803013440" observedRunningTime="2025-09-29 19:12:24.716121939 +0000 UTC m=+186.363911271" watchObservedRunningTime="2025-09-29 19:12:24.717760061 +0000 UTC m=+186.365549393" Sep 29 19:12:27 crc kubenswrapper[4741]: I0929 19:12:27.057297 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:12:27 crc kubenswrapper[4741]: I0929 19:12:27.057630 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:12:27 crc kubenswrapper[4741]: I0929 19:12:27.204001 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:12:31 crc kubenswrapper[4741]: I0929 19:12:31.738637 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:12:31 crc kubenswrapper[4741]: I0929 19:12:31.739027 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:12:34 crc kubenswrapper[4741]: I0929 19:12:34.724053 4741 generic.go:334] "Generic (PLEG): container finished" podID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerID="9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b" exitCode=0 Sep 29 19:12:34 crc kubenswrapper[4741]: I0929 19:12:34.724118 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdvm7" event={"ID":"bbbbb850-78fa-47f0-8339-24148d9c610e","Type":"ContainerDied","Data":"9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b"} Sep 29 19:12:35 crc kubenswrapper[4741]: I0929 19:12:35.730473 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdvm7" event={"ID":"bbbbb850-78fa-47f0-8339-24148d9c610e","Type":"ContainerStarted","Data":"6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d"} Sep 29 19:12:35 crc kubenswrapper[4741]: I0929 19:12:35.749539 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rdvm7" podStartSLOduration=1.837711293 podStartE2EDuration="46.749523308s" podCreationTimestamp="2025-09-29 19:11:49 +0000 UTC" firstStartedPulling="2025-09-29 19:11:50.421311582 +0000 UTC m=+152.069100914" lastFinishedPulling="2025-09-29 19:12:35.333123597 +0000 UTC m=+196.980912929" observedRunningTime="2025-09-29 19:12:35.7477045 +0000 UTC m=+197.395493882" watchObservedRunningTime="2025-09-29 19:12:35.749523308 +0000 UTC m=+197.397312640" Sep 29 19:12:36 crc kubenswrapper[4741]: I0929 19:12:36.736680 4741 generic.go:334] "Generic (PLEG): container finished" podID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerID="14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4" exitCode=0 Sep 29 19:12:36 crc kubenswrapper[4741]: I0929 19:12:36.736775 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf8jv" event={"ID":"b13e515d-1615-4338-b8b3-1b08e4d3faf4","Type":"ContainerDied","Data":"14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4"} Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.106624 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.742844 4741 generic.go:334] "Generic (PLEG): container finished" podID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerID="4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e" exitCode=0 Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.742911 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdnwf" event={"ID":"8e4390f3-92f6-4639-81ad-c1d11073309e","Type":"ContainerDied","Data":"4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e"} Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.744664 4741 generic.go:334] "Generic (PLEG): container finished" podID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerID="723f531b35571bd73973e65853844683582ce74854dd892e051558ad0f171eac" exitCode=0 Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.744716 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c26bp" event={"ID":"c9bd836e-9017-45c5-9de1-d8bcba956088","Type":"ContainerDied","Data":"723f531b35571bd73973e65853844683582ce74854dd892e051558ad0f171eac"} Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.746424 4741 generic.go:334] "Generic (PLEG): container finished" podID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerID="b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d" exitCode=0 Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.746621 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vmvz" event={"ID":"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9","Type":"ContainerDied","Data":"b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d"} Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.749309 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf8jv" event={"ID":"b13e515d-1615-4338-b8b3-1b08e4d3faf4","Type":"ContainerStarted","Data":"747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d"} Sep 29 19:12:37 crc kubenswrapper[4741]: I0929 19:12:37.801533 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bf8jv" podStartSLOduration=2.787451148 podStartE2EDuration="49.801514868s" podCreationTimestamp="2025-09-29 19:11:48 +0000 UTC" firstStartedPulling="2025-09-29 19:11:50.428055499 +0000 UTC m=+152.075844831" lastFinishedPulling="2025-09-29 19:12:37.442119219 +0000 UTC m=+199.089908551" observedRunningTime="2025-09-29 19:12:37.798710329 +0000 UTC m=+199.446499661" watchObservedRunningTime="2025-09-29 19:12:37.801514868 +0000 UTC m=+199.449304200" Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.757069 4741 generic.go:334] "Generic (PLEG): container finished" podID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerID="780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0" exitCode=0 Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.757153 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ggfh" event={"ID":"3a5e870e-8a78-4835-b453-764cae6e4b41","Type":"ContainerDied","Data":"780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0"} Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.760059 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vmvz" event={"ID":"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9","Type":"ContainerStarted","Data":"462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c"} Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.762672 4741 generic.go:334] "Generic (PLEG): container finished" podID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerID="6591847a4bd829213f8292ce1506b707868e2e08f67cff424dd96b606ef349ca" exitCode=0 Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.762710 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kcsrb" event={"ID":"e92734a4-fea9-42b6-badf-dc94c29db4fc","Type":"ContainerDied","Data":"6591847a4bd829213f8292ce1506b707868e2e08f67cff424dd96b606ef349ca"} Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.765616 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdnwf" event={"ID":"8e4390f3-92f6-4639-81ad-c1d11073309e","Type":"ContainerStarted","Data":"f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9"} Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.771093 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.771126 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c26bp" event={"ID":"c9bd836e-9017-45c5-9de1-d8bcba956088","Type":"ContainerStarted","Data":"7116b971ecc2be02e630106b46ead40e546b84f444e211d4787d3b410bc58022"} Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.771146 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.807375 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c26bp" podStartSLOduration=1.993924114 podStartE2EDuration="52.807330051s" podCreationTimestamp="2025-09-29 19:11:46 +0000 UTC" firstStartedPulling="2025-09-29 19:11:47.344613956 +0000 UTC m=+148.992403288" lastFinishedPulling="2025-09-29 19:12:38.158019893 +0000 UTC m=+199.805809225" observedRunningTime="2025-09-29 19:12:38.804499871 +0000 UTC m=+200.452289203" watchObservedRunningTime="2025-09-29 19:12:38.807330051 +0000 UTC m=+200.455119393" Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.827878 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jdnwf" podStartSLOduration=2.985065849 podStartE2EDuration="49.827860362s" podCreationTimestamp="2025-09-29 19:11:49 +0000 UTC" firstStartedPulling="2025-09-29 19:11:51.470754441 +0000 UTC m=+153.118543773" lastFinishedPulling="2025-09-29 19:12:38.313548934 +0000 UTC m=+199.961338286" observedRunningTime="2025-09-29 19:12:38.825308311 +0000 UTC m=+200.473097643" watchObservedRunningTime="2025-09-29 19:12:38.827860362 +0000 UTC m=+200.475649694" Sep 29 19:12:38 crc kubenswrapper[4741]: I0929 19:12:38.846931 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2vmvz" podStartSLOduration=1.869569581 podStartE2EDuration="50.846914907s" podCreationTimestamp="2025-09-29 19:11:48 +0000 UTC" firstStartedPulling="2025-09-29 19:11:49.409876444 +0000 UTC m=+151.057665776" lastFinishedPulling="2025-09-29 19:12:38.38722178 +0000 UTC m=+200.035011102" observedRunningTime="2025-09-29 19:12:38.846274346 +0000 UTC m=+200.494063678" watchObservedRunningTime="2025-09-29 19:12:38.846914907 +0000 UTC m=+200.494704239" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.161970 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.162027 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.206840 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.754123 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.754472 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.779289 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kcsrb" event={"ID":"e92734a4-fea9-42b6-badf-dc94c29db4fc","Type":"ContainerStarted","Data":"0c36febf77e49538a39218224ce24f0835247300b12e21266ed524b41bb3b639"} Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.783423 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ggfh" event={"ID":"3a5e870e-8a78-4835-b453-764cae6e4b41","Type":"ContainerStarted","Data":"8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4"} Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.803011 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kcsrb" podStartSLOduration=1.952310585 podStartE2EDuration="53.802995242s" podCreationTimestamp="2025-09-29 19:11:46 +0000 UTC" firstStartedPulling="2025-09-29 19:11:47.368455052 +0000 UTC m=+149.016244384" lastFinishedPulling="2025-09-29 19:12:39.219139709 +0000 UTC m=+200.866929041" observedRunningTime="2025-09-29 19:12:39.801434932 +0000 UTC m=+201.449224274" watchObservedRunningTime="2025-09-29 19:12:39.802995242 +0000 UTC m=+201.450784574" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.824516 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6ggfh" podStartSLOduration=2.911564196 podStartE2EDuration="53.824496143s" podCreationTimestamp="2025-09-29 19:11:46 +0000 UTC" firstStartedPulling="2025-09-29 19:11:48.39889646 +0000 UTC m=+150.046685782" lastFinishedPulling="2025-09-29 19:12:39.311828397 +0000 UTC m=+200.959617729" observedRunningTime="2025-09-29 19:12:39.821718995 +0000 UTC m=+201.469508327" watchObservedRunningTime="2025-09-29 19:12:39.824496143 +0000 UTC m=+201.472285475" Sep 29 19:12:39 crc kubenswrapper[4741]: I0929 19:12:39.826145 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-2vmvz" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="registry-server" probeResult="failure" output=< Sep 29 19:12:39 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 19:12:39 crc kubenswrapper[4741]: > Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.157032 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.157296 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.519736 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lfcss"] Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.520248 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lfcss" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="registry-server" containerID="cri-o://d06c4023687ee93b59c35ba59f85be2f0ea5cc762e5d0b64e4112553f2dfbe99" gracePeriod=2 Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.792193 4741 generic.go:334] "Generic (PLEG): container finished" podID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerID="d06c4023687ee93b59c35ba59f85be2f0ea5cc762e5d0b64e4112553f2dfbe99" exitCode=0 Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.793243 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfcss" event={"ID":"84b44513-b985-42e9-ba2c-1c3af7c6ebd1","Type":"ContainerDied","Data":"d06c4023687ee93b59c35ba59f85be2f0ea5cc762e5d0b64e4112553f2dfbe99"} Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.793325 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rdvm7" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="registry-server" probeResult="failure" output=< Sep 29 19:12:40 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 19:12:40 crc kubenswrapper[4741]: > Sep 29 19:12:40 crc kubenswrapper[4741]: I0929 19:12:40.915519 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.007139 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-catalog-content\") pod \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.007217 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-utilities\") pod \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.007256 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwm6v\" (UniqueName: \"kubernetes.io/projected/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-kube-api-access-xwm6v\") pod \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\" (UID: \"84b44513-b985-42e9-ba2c-1c3af7c6ebd1\") " Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.007988 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-utilities" (OuterVolumeSpecName: "utilities") pod "84b44513-b985-42e9-ba2c-1c3af7c6ebd1" (UID: "84b44513-b985-42e9-ba2c-1c3af7c6ebd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.018583 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-kube-api-access-xwm6v" (OuterVolumeSpecName: "kube-api-access-xwm6v") pod "84b44513-b985-42e9-ba2c-1c3af7c6ebd1" (UID: "84b44513-b985-42e9-ba2c-1c3af7c6ebd1"). InnerVolumeSpecName "kube-api-access-xwm6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.054897 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84b44513-b985-42e9-ba2c-1c3af7c6ebd1" (UID: "84b44513-b985-42e9-ba2c-1c3af7c6ebd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.108171 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.108203 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.108213 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwm6v\" (UniqueName: \"kubernetes.io/projected/84b44513-b985-42e9-ba2c-1c3af7c6ebd1-kube-api-access-xwm6v\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.190337 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jdnwf" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="registry-server" probeResult="failure" output=< Sep 29 19:12:41 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 19:12:41 crc kubenswrapper[4741]: > Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.799119 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfcss" event={"ID":"84b44513-b985-42e9-ba2c-1c3af7c6ebd1","Type":"ContainerDied","Data":"a3abb031b69717f8e1ea5a7c1e49af245e7140f48bbb6cc08ad232671038d451"} Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.799152 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfcss" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.799412 4741 scope.go:117] "RemoveContainer" containerID="d06c4023687ee93b59c35ba59f85be2f0ea5cc762e5d0b64e4112553f2dfbe99" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.814167 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lfcss"] Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.818754 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lfcss"] Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.819024 4741 scope.go:117] "RemoveContainer" containerID="443a784b4511ac91e0a98c6e73658191492ae729f7a1211245c601544e355022" Sep 29 19:12:41 crc kubenswrapper[4741]: I0929 19:12:41.840678 4741 scope.go:117] "RemoveContainer" containerID="d9054f9e7f8ccd4e483bfbb507f845674d9b28f02553e9eb743bbcb5ae39a42f" Sep 29 19:12:43 crc kubenswrapper[4741]: I0929 19:12:43.092199 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" path="/var/lib/kubelet/pods/84b44513-b985-42e9-ba2c-1c3af7c6ebd1/volumes" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.593482 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.594100 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.635987 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.783639 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.783690 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.822817 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.862357 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:12:46 crc kubenswrapper[4741]: I0929 19:12:46.872661 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:12:47 crc kubenswrapper[4741]: I0929 19:12:47.164741 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:12:47 crc kubenswrapper[4741]: I0929 19:12:47.164816 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:12:47 crc kubenswrapper[4741]: I0929 19:12:47.206679 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:12:47 crc kubenswrapper[4741]: I0929 19:12:47.867691 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:12:48 crc kubenswrapper[4741]: I0929 19:12:48.652720 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ggfh"] Sep 29 19:12:48 crc kubenswrapper[4741]: I0929 19:12:48.817787 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:12:48 crc kubenswrapper[4741]: I0929 19:12:48.859235 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:12:49 crc kubenswrapper[4741]: I0929 19:12:49.213425 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:12:49 crc kubenswrapper[4741]: I0929 19:12:49.792252 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:12:49 crc kubenswrapper[4741]: I0929 19:12:49.836299 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:12:49 crc kubenswrapper[4741]: I0929 19:12:49.840856 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6ggfh" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="registry-server" containerID="cri-o://8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4" gracePeriod=2 Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.192323 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.233968 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.812686 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.845365 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d6d6\" (UniqueName: \"kubernetes.io/projected/3a5e870e-8a78-4835-b453-764cae6e4b41-kube-api-access-4d6d6\") pod \"3a5e870e-8a78-4835-b453-764cae6e4b41\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.845776 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-catalog-content\") pod \"3a5e870e-8a78-4835-b453-764cae6e4b41\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.845952 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-utilities\") pod \"3a5e870e-8a78-4835-b453-764cae6e4b41\" (UID: \"3a5e870e-8a78-4835-b453-764cae6e4b41\") " Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.847811 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-utilities" (OuterVolumeSpecName: "utilities") pod "3a5e870e-8a78-4835-b453-764cae6e4b41" (UID: "3a5e870e-8a78-4835-b453-764cae6e4b41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.849981 4741 generic.go:334] "Generic (PLEG): container finished" podID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerID="8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4" exitCode=0 Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.850743 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ggfh" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.851418 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ggfh" event={"ID":"3a5e870e-8a78-4835-b453-764cae6e4b41","Type":"ContainerDied","Data":"8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4"} Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.851528 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ggfh" event={"ID":"3a5e870e-8a78-4835-b453-764cae6e4b41","Type":"ContainerDied","Data":"f0e11efb117f5d56943ee5472c4a16b5bce26e7be2c5736e09f573bb22a10615"} Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.851604 4741 scope.go:117] "RemoveContainer" containerID="8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.854536 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a5e870e-8a78-4835-b453-764cae6e4b41-kube-api-access-4d6d6" (OuterVolumeSpecName: "kube-api-access-4d6d6") pod "3a5e870e-8a78-4835-b453-764cae6e4b41" (UID: "3a5e870e-8a78-4835-b453-764cae6e4b41"). InnerVolumeSpecName "kube-api-access-4d6d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.880433 4741 scope.go:117] "RemoveContainer" containerID="780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.897506 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a5e870e-8a78-4835-b453-764cae6e4b41" (UID: "3a5e870e-8a78-4835-b453-764cae6e4b41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.898082 4741 scope.go:117] "RemoveContainer" containerID="b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.919624 4741 scope.go:117] "RemoveContainer" containerID="8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4" Sep 29 19:12:50 crc kubenswrapper[4741]: E0929 19:12:50.920189 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4\": container with ID starting with 8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4 not found: ID does not exist" containerID="8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.920232 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4"} err="failed to get container status \"8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4\": rpc error: code = NotFound desc = could not find container \"8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4\": container with ID starting with 8f66338f0e3f14320a6f9ea821bc8968171898d41185f31b66630b8f1c62a5a4 not found: ID does not exist" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.920284 4741 scope.go:117] "RemoveContainer" containerID="780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0" Sep 29 19:12:50 crc kubenswrapper[4741]: E0929 19:12:50.920625 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0\": container with ID starting with 780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0 not found: ID does not exist" containerID="780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.920672 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0"} err="failed to get container status \"780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0\": rpc error: code = NotFound desc = could not find container \"780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0\": container with ID starting with 780dd9d97acc93edc1d54d2a0ba2dfbfc0c29202a09259bf8d95b7db6e9834f0 not found: ID does not exist" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.920706 4741 scope.go:117] "RemoveContainer" containerID="b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27" Sep 29 19:12:50 crc kubenswrapper[4741]: E0929 19:12:50.921011 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27\": container with ID starting with b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27 not found: ID does not exist" containerID="b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.921036 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27"} err="failed to get container status \"b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27\": rpc error: code = NotFound desc = could not find container \"b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27\": container with ID starting with b0aef47fa0d5ad70c3718d89202e55e6df7f9d43aace89675b1b4a52c4a0dc27 not found: ID does not exist" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.947161 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.947191 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d6d6\" (UniqueName: \"kubernetes.io/projected/3a5e870e-8a78-4835-b453-764cae6e4b41-kube-api-access-4d6d6\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:50 crc kubenswrapper[4741]: I0929 19:12:50.947204 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a5e870e-8a78-4835-b453-764cae6e4b41-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.165578 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ggfh"] Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.168331 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6ggfh"] Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.251325 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf8jv"] Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.252094 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bf8jv" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="registry-server" containerID="cri-o://747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d" gracePeriod=2 Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.573812 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.655183 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-catalog-content\") pod \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.663147 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-utilities\") pod \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.663279 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjnwm\" (UniqueName: \"kubernetes.io/projected/b13e515d-1615-4338-b8b3-1b08e4d3faf4-kube-api-access-jjnwm\") pod \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\" (UID: \"b13e515d-1615-4338-b8b3-1b08e4d3faf4\") " Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.663819 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-utilities" (OuterVolumeSpecName: "utilities") pod "b13e515d-1615-4338-b8b3-1b08e4d3faf4" (UID: "b13e515d-1615-4338-b8b3-1b08e4d3faf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.668922 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b13e515d-1615-4338-b8b3-1b08e4d3faf4" (UID: "b13e515d-1615-4338-b8b3-1b08e4d3faf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.669807 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13e515d-1615-4338-b8b3-1b08e4d3faf4-kube-api-access-jjnwm" (OuterVolumeSpecName: "kube-api-access-jjnwm") pod "b13e515d-1615-4338-b8b3-1b08e4d3faf4" (UID: "b13e515d-1615-4338-b8b3-1b08e4d3faf4"). InnerVolumeSpecName "kube-api-access-jjnwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.764938 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.764975 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjnwm\" (UniqueName: \"kubernetes.io/projected/b13e515d-1615-4338-b8b3-1b08e4d3faf4-kube-api-access-jjnwm\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.764986 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13e515d-1615-4338-b8b3-1b08e4d3faf4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.857680 4741 generic.go:334] "Generic (PLEG): container finished" podID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerID="747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d" exitCode=0 Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.857743 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf8jv" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.857971 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf8jv" event={"ID":"b13e515d-1615-4338-b8b3-1b08e4d3faf4","Type":"ContainerDied","Data":"747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d"} Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.858044 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf8jv" event={"ID":"b13e515d-1615-4338-b8b3-1b08e4d3faf4","Type":"ContainerDied","Data":"5fb45479d4f59bf84f3d11e55e871febc2d8623331bdefab496440a069d4d989"} Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.858064 4741 scope.go:117] "RemoveContainer" containerID="747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.888787 4741 scope.go:117] "RemoveContainer" containerID="14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.889595 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf8jv"] Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.892323 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf8jv"] Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.918462 4741 scope.go:117] "RemoveContainer" containerID="15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.933902 4741 scope.go:117] "RemoveContainer" containerID="747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d" Sep 29 19:12:51 crc kubenswrapper[4741]: E0929 19:12:51.934229 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d\": container with ID starting with 747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d not found: ID does not exist" containerID="747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.934265 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d"} err="failed to get container status \"747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d\": rpc error: code = NotFound desc = could not find container \"747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d\": container with ID starting with 747fd4cb5b714418f4fb6a1e088fc5379dc99fd38090f02e7f0f9e992447084d not found: ID does not exist" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.934288 4741 scope.go:117] "RemoveContainer" containerID="14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4" Sep 29 19:12:51 crc kubenswrapper[4741]: E0929 19:12:51.934633 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4\": container with ID starting with 14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4 not found: ID does not exist" containerID="14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.934691 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4"} err="failed to get container status \"14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4\": rpc error: code = NotFound desc = could not find container \"14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4\": container with ID starting with 14319a156b5334a663efc355a551d93281fae9809ec1d134080612483d8ebde4 not found: ID does not exist" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.934731 4741 scope.go:117] "RemoveContainer" containerID="15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946" Sep 29 19:12:51 crc kubenswrapper[4741]: E0929 19:12:51.935901 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946\": container with ID starting with 15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946 not found: ID does not exist" containerID="15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946" Sep 29 19:12:51 crc kubenswrapper[4741]: I0929 19:12:51.935933 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946"} err="failed to get container status \"15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946\": rpc error: code = NotFound desc = could not find container \"15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946\": container with ID starting with 15e7f5e5810d80d9cb04961d7a9ae5b19e20ac82717e6697b9c7a30bb83ce946 not found: ID does not exist" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.092123 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" path="/var/lib/kubelet/pods/3a5e870e-8a78-4835-b453-764cae6e4b41/volumes" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.092820 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" path="/var/lib/kubelet/pods/b13e515d-1615-4338-b8b3-1b08e4d3faf4/volumes" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.454251 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdnwf"] Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.454820 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jdnwf" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="registry-server" containerID="cri-o://f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9" gracePeriod=2 Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.819543 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.889435 4741 generic.go:334] "Generic (PLEG): container finished" podID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerID="f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9" exitCode=0 Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.889477 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdnwf" event={"ID":"8e4390f3-92f6-4639-81ad-c1d11073309e","Type":"ContainerDied","Data":"f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9"} Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.889508 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdnwf" event={"ID":"8e4390f3-92f6-4639-81ad-c1d11073309e","Type":"ContainerDied","Data":"70dbe3e201770588765f81086f80ca7efdaea2fe37bd570afa1bcf3d45405f39"} Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.889526 4741 scope.go:117] "RemoveContainer" containerID="f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.889536 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdnwf" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.892739 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-catalog-content\") pod \"8e4390f3-92f6-4639-81ad-c1d11073309e\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.892803 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l24sk\" (UniqueName: \"kubernetes.io/projected/8e4390f3-92f6-4639-81ad-c1d11073309e-kube-api-access-l24sk\") pod \"8e4390f3-92f6-4639-81ad-c1d11073309e\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.892841 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-utilities\") pod \"8e4390f3-92f6-4639-81ad-c1d11073309e\" (UID: \"8e4390f3-92f6-4639-81ad-c1d11073309e\") " Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.893794 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-utilities" (OuterVolumeSpecName: "utilities") pod "8e4390f3-92f6-4639-81ad-c1d11073309e" (UID: "8e4390f3-92f6-4639-81ad-c1d11073309e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.904577 4741 scope.go:117] "RemoveContainer" containerID="4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.904616 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e4390f3-92f6-4639-81ad-c1d11073309e-kube-api-access-l24sk" (OuterVolumeSpecName: "kube-api-access-l24sk") pod "8e4390f3-92f6-4639-81ad-c1d11073309e" (UID: "8e4390f3-92f6-4639-81ad-c1d11073309e"). InnerVolumeSpecName "kube-api-access-l24sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.919623 4741 scope.go:117] "RemoveContainer" containerID="01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.934706 4741 scope.go:117] "RemoveContainer" containerID="f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9" Sep 29 19:12:53 crc kubenswrapper[4741]: E0929 19:12:53.935202 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9\": container with ID starting with f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9 not found: ID does not exist" containerID="f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.935264 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9"} err="failed to get container status \"f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9\": rpc error: code = NotFound desc = could not find container \"f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9\": container with ID starting with f0a97c3762200a4af38139406eb64c4226495d53fe257ed9b917b03f061db6e9 not found: ID does not exist" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.935288 4741 scope.go:117] "RemoveContainer" containerID="4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e" Sep 29 19:12:53 crc kubenswrapper[4741]: E0929 19:12:53.935714 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e\": container with ID starting with 4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e not found: ID does not exist" containerID="4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.935736 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e"} err="failed to get container status \"4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e\": rpc error: code = NotFound desc = could not find container \"4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e\": container with ID starting with 4d6ee466fe28b564e12606fdf638b73570082ecb07cc5077a1e38756c4fc916e not found: ID does not exist" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.935752 4741 scope.go:117] "RemoveContainer" containerID="01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d" Sep 29 19:12:53 crc kubenswrapper[4741]: E0929 19:12:53.935990 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d\": container with ID starting with 01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d not found: ID does not exist" containerID="01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.936016 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d"} err="failed to get container status \"01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d\": rpc error: code = NotFound desc = could not find container \"01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d\": container with ID starting with 01053f2a0a08906aaa69a387a784725c5050ea32ae1ffce6fe7d2aaed0f7a18d not found: ID does not exist" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.983332 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e4390f3-92f6-4639-81ad-c1d11073309e" (UID: "8e4390f3-92f6-4639-81ad-c1d11073309e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.994050 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.994083 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l24sk\" (UniqueName: \"kubernetes.io/projected/8e4390f3-92f6-4639-81ad-c1d11073309e-kube-api-access-l24sk\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:53 crc kubenswrapper[4741]: I0929 19:12:53.994097 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e4390f3-92f6-4639-81ad-c1d11073309e-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:12:54 crc kubenswrapper[4741]: I0929 19:12:54.216092 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdnwf"] Sep 29 19:12:54 crc kubenswrapper[4741]: I0929 19:12:54.219527 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jdnwf"] Sep 29 19:12:55 crc kubenswrapper[4741]: I0929 19:12:55.091781 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" path="/var/lib/kubelet/pods/8e4390f3-92f6-4639-81ad-c1d11073309e/volumes" Sep 29 19:12:59 crc kubenswrapper[4741]: I0929 19:12:59.045533 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzlbd"] Sep 29 19:13:01 crc kubenswrapper[4741]: I0929 19:13:01.739535 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:13:01 crc kubenswrapper[4741]: I0929 19:13:01.739926 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:13:01 crc kubenswrapper[4741]: I0929 19:13:01.739988 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:13:01 crc kubenswrapper[4741]: I0929 19:13:01.740724 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:13:01 crc kubenswrapper[4741]: I0929 19:13:01.740811 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2" gracePeriod=600 Sep 29 19:13:01 crc kubenswrapper[4741]: I0929 19:13:01.950255 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2" exitCode=0 Sep 29 19:13:01 crc kubenswrapper[4741]: I0929 19:13:01.950338 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2"} Sep 29 19:13:02 crc kubenswrapper[4741]: I0929 19:13:02.963644 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"7d272b5f97846d5ba906fda7647c908fb192dc3a478a94f6b285f09880ecc623"} Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.071037 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" podUID="bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" containerName="oauth-openshift" containerID="cri-o://b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e" gracePeriod=15 Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.416113 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450540 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-985c66b4-zbct9"] Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450788 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450804 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450818 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450826 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450838 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450847 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450858 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450865 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450879 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7e827b-32c0-4af9-8f81-34a549a3c871" containerName="pruner" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450886 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7e827b-32c0-4af9-8f81-34a549a3c871" containerName="pruner" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450897 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" containerName="oauth-openshift" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450906 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" containerName="oauth-openshift" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450918 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450926 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450936 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450944 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450953 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450961 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="extract-content" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450972 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450979 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.450990 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.450998 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.451009 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451019 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="extract-utilities" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.451029 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451037 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: E0929 19:13:24.451054 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451062 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451203 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" containerName="oauth-openshift" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451218 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13e515d-1615-4338-b8b3-1b08e4d3faf4" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451227 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a5e870e-8a78-4835-b453-764cae6e4b41" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451236 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b44513-b985-42e9-ba2c-1c3af7c6ebd1" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451252 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e7e827b-32c0-4af9-8f81-34a549a3c871" containerName="pruner" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451264 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4390f3-92f6-4639-81ad-c1d11073309e" containerName="registry-server" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.451692 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.459628 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-985c66b4-zbct9"] Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.469822 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-trusted-ca-bundle\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.469865 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-cliconfig\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.469889 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-service-ca\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.469907 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-session\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.469933 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-dir\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.469951 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-idp-0-file-data\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.469974 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-router-certs\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.470002 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-login\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.470022 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-provider-selection\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.470062 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-policies\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.470097 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-serving-cert\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.470116 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-ocp-branding-template\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.470136 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6lgr\" (UniqueName: \"kubernetes.io/projected/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-kube-api-access-j6lgr\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.470152 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-error\") pod \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\" (UID: \"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6\") " Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.472314 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.472871 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.472950 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.473313 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.473348 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.476374 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.477786 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.479090 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.491596 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.491858 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.492060 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.492425 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.492731 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-kube-api-access-j6lgr" (OuterVolumeSpecName: "kube-api-access-j6lgr") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "kube-api-access-j6lgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.494369 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" (UID: "bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571273 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571317 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-login\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571336 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf4j4\" (UniqueName: \"kubernetes.io/projected/67fe275e-24e1-4cac-b10b-204819740151-kube-api-access-sf4j4\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571361 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-error\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571375 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571453 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-serving-cert\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571481 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-session\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571502 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571535 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-audit-policies\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571573 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-cliconfig\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571593 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-router-certs\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571614 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571639 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67fe275e-24e1-4cac-b10b-204819740151-audit-dir\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571661 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-service-ca\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571692 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571703 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571714 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6lgr\" (UniqueName: \"kubernetes.io/projected/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-kube-api-access-j6lgr\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571723 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571733 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571742 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571751 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571761 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571771 4741 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-dir\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571780 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571789 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571798 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571808 4741 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.571818 4741 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673210 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673260 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-login\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673282 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf4j4\" (UniqueName: \"kubernetes.io/projected/67fe275e-24e1-4cac-b10b-204819740151-kube-api-access-sf4j4\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673302 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-error\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673317 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673357 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-session\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673380 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-serving-cert\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673417 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673628 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-audit-policies\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673650 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-router-certs\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673664 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-cliconfig\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673684 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673712 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67fe275e-24e1-4cac-b10b-204819740151-audit-dir\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.673728 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-service-ca\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.674251 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67fe275e-24e1-4cac-b10b-204819740151-audit-dir\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.674557 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-service-ca\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.674598 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-audit-policies\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.675475 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-cliconfig\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.675485 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.677854 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-session\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.677895 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-serving-cert\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.678449 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.678641 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.678736 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-login\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.678882 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.679582 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-system-router-certs\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.679606 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/67fe275e-24e1-4cac-b10b-204819740151-v4-0-config-user-template-error\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.689955 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf4j4\" (UniqueName: \"kubernetes.io/projected/67fe275e-24e1-4cac-b10b-204819740151-kube-api-access-sf4j4\") pod \"oauth-openshift-985c66b4-zbct9\" (UID: \"67fe275e-24e1-4cac-b10b-204819740151\") " pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:24 crc kubenswrapper[4741]: I0929 19:13:24.772267 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.085112 4741 generic.go:334] "Generic (PLEG): container finished" podID="bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" containerID="b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e" exitCode=0 Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.085340 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.102580 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" event={"ID":"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6","Type":"ContainerDied","Data":"b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e"} Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.102620 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzlbd" event={"ID":"bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6","Type":"ContainerDied","Data":"e4bbf68e9561016712f397ac6841b4cfc0062e71b0e1fc66e2b4f579385ea4a8"} Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.102643 4741 scope.go:117] "RemoveContainer" containerID="b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e" Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.126862 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzlbd"] Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.130558 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzlbd"] Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.138737 4741 scope.go:117] "RemoveContainer" containerID="b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e" Sep 29 19:13:25 crc kubenswrapper[4741]: E0929 19:13:25.139179 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e\": container with ID starting with b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e not found: ID does not exist" containerID="b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e" Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.139219 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e"} err="failed to get container status \"b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e\": rpc error: code = NotFound desc = could not find container \"b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e\": container with ID starting with b51e3cfa8236e2efd6b61c4836d96437bb06ce52d204575023cc8fa7046cd27e not found: ID does not exist" Sep 29 19:13:25 crc kubenswrapper[4741]: I0929 19:13:25.184045 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-985c66b4-zbct9"] Sep 29 19:13:25 crc kubenswrapper[4741]: W0929 19:13:25.191729 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67fe275e_24e1_4cac_b10b_204819740151.slice/crio-5e8aacf87ba9345201480a02e1e1b7b21cd71bb8b3fdf5f031b62beaf9fa90dd WatchSource:0}: Error finding container 5e8aacf87ba9345201480a02e1e1b7b21cd71bb8b3fdf5f031b62beaf9fa90dd: Status 404 returned error can't find the container with id 5e8aacf87ba9345201480a02e1e1b7b21cd71bb8b3fdf5f031b62beaf9fa90dd Sep 29 19:13:26 crc kubenswrapper[4741]: I0929 19:13:26.091346 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" event={"ID":"67fe275e-24e1-4cac-b10b-204819740151","Type":"ContainerStarted","Data":"96ae41bba5890d94aae00da722e4a4e1861ec66f42597a811097e1a0e77b6f36"} Sep 29 19:13:26 crc kubenswrapper[4741]: I0929 19:13:26.091716 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:26 crc kubenswrapper[4741]: I0929 19:13:26.091730 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" event={"ID":"67fe275e-24e1-4cac-b10b-204819740151","Type":"ContainerStarted","Data":"5e8aacf87ba9345201480a02e1e1b7b21cd71bb8b3fdf5f031b62beaf9fa90dd"} Sep 29 19:13:26 crc kubenswrapper[4741]: I0929 19:13:26.096949 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" Sep 29 19:13:26 crc kubenswrapper[4741]: I0929 19:13:26.117990 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-985c66b4-zbct9" podStartSLOduration=27.117971741 podStartE2EDuration="27.117971741s" podCreationTimestamp="2025-09-29 19:12:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:13:26.11475582 +0000 UTC m=+247.762545142" watchObservedRunningTime="2025-09-29 19:13:26.117971741 +0000 UTC m=+247.765761093" Sep 29 19:13:27 crc kubenswrapper[4741]: I0929 19:13:27.091911 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6" path="/var/lib/kubelet/pods/bf794dcd-0fcf-42e3-a45e-1b0a9dc152e6/volumes" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.047312 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kcsrb"] Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.048076 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kcsrb" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="registry-server" containerID="cri-o://0c36febf77e49538a39218224ce24f0835247300b12e21266ed524b41bb3b639" gracePeriod=30 Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.054501 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c26bp"] Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.054856 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c26bp" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="registry-server" containerID="cri-o://7116b971ecc2be02e630106b46ead40e546b84f444e211d4787d3b410bc58022" gracePeriod=30 Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.069690 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fjfzz"] Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.069967 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" podUID="2d7a0ce6-fda6-4100-801b-a7170a4d5e93" containerName="marketplace-operator" containerID="cri-o://68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63" gracePeriod=30 Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.076938 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vmvz"] Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.077190 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2vmvz" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="registry-server" containerID="cri-o://462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c" gracePeriod=30 Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.107142 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdvm7"] Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.107635 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rdvm7" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="registry-server" containerID="cri-o://6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d" gracePeriod=30 Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.113528 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fhs5q"] Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.114476 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.115266 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fhs5q"] Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.175270 4741 generic.go:334] "Generic (PLEG): container finished" podID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerID="0c36febf77e49538a39218224ce24f0835247300b12e21266ed524b41bb3b639" exitCode=0 Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.175332 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kcsrb" event={"ID":"e92734a4-fea9-42b6-badf-dc94c29db4fc","Type":"ContainerDied","Data":"0c36febf77e49538a39218224ce24f0835247300b12e21266ed524b41bb3b639"} Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.177902 4741 generic.go:334] "Generic (PLEG): container finished" podID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerID="7116b971ecc2be02e630106b46ead40e546b84f444e211d4787d3b410bc58022" exitCode=0 Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.177925 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c26bp" event={"ID":"c9bd836e-9017-45c5-9de1-d8bcba956088","Type":"ContainerDied","Data":"7116b971ecc2be02e630106b46ead40e546b84f444e211d4787d3b410bc58022"} Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.233484 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw9zc\" (UniqueName: \"kubernetes.io/projected/674ba05b-0612-46ca-b5b6-a91eb3950d27-kube-api-access-mw9zc\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.233564 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/674ba05b-0612-46ca-b5b6-a91eb3950d27-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.233611 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/674ba05b-0612-46ca-b5b6-a91eb3950d27-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.335941 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/674ba05b-0612-46ca-b5b6-a91eb3950d27-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.336322 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw9zc\" (UniqueName: \"kubernetes.io/projected/674ba05b-0612-46ca-b5b6-a91eb3950d27-kube-api-access-mw9zc\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.336424 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/674ba05b-0612-46ca-b5b6-a91eb3950d27-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.338037 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/674ba05b-0612-46ca-b5b6-a91eb3950d27-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.342673 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/674ba05b-0612-46ca-b5b6-a91eb3950d27-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.356238 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw9zc\" (UniqueName: \"kubernetes.io/projected/674ba05b-0612-46ca-b5b6-a91eb3950d27-kube-api-access-mw9zc\") pod \"marketplace-operator-79b997595-fhs5q\" (UID: \"674ba05b-0612-46ca-b5b6-a91eb3950d27\") " pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.525947 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.526482 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.546579 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.552367 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.554633 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.566939 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641214 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-catalog-content\") pod \"e92734a4-fea9-42b6-badf-dc94c29db4fc\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641258 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-catalog-content\") pod \"bbbbb850-78fa-47f0-8339-24148d9c610e\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641275 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-utilities\") pod \"bbbbb850-78fa-47f0-8339-24148d9c610e\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641298 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-utilities\") pod \"c9bd836e-9017-45c5-9de1-d8bcba956088\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641313 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-utilities\") pod \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641342 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-operator-metrics\") pod \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641368 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcbvz\" (UniqueName: \"kubernetes.io/projected/e92734a4-fea9-42b6-badf-dc94c29db4fc-kube-api-access-lcbvz\") pod \"e92734a4-fea9-42b6-badf-dc94c29db4fc\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641407 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-trusted-ca\") pod \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641670 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-catalog-content\") pod \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641717 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndfnk\" (UniqueName: \"kubernetes.io/projected/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-kube-api-access-ndfnk\") pod \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\" (UID: \"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641751 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-utilities\") pod \"e92734a4-fea9-42b6-badf-dc94c29db4fc\" (UID: \"e92734a4-fea9-42b6-badf-dc94c29db4fc\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641806 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-865jv\" (UniqueName: \"kubernetes.io/projected/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-kube-api-access-865jv\") pod \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\" (UID: \"2d7a0ce6-fda6-4100-801b-a7170a4d5e93\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641828 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gtk9\" (UniqueName: \"kubernetes.io/projected/bbbbb850-78fa-47f0-8339-24148d9c610e-kube-api-access-9gtk9\") pod \"bbbbb850-78fa-47f0-8339-24148d9c610e\" (UID: \"bbbbb850-78fa-47f0-8339-24148d9c610e\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641845 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-catalog-content\") pod \"c9bd836e-9017-45c5-9de1-d8bcba956088\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.641865 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lrb8\" (UniqueName: \"kubernetes.io/projected/c9bd836e-9017-45c5-9de1-d8bcba956088-kube-api-access-6lrb8\") pod \"c9bd836e-9017-45c5-9de1-d8bcba956088\" (UID: \"c9bd836e-9017-45c5-9de1-d8bcba956088\") " Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.642246 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2d7a0ce6-fda6-4100-801b-a7170a4d5e93" (UID: "2d7a0ce6-fda6-4100-801b-a7170a4d5e93"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.642778 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-utilities" (OuterVolumeSpecName: "utilities") pod "c9bd836e-9017-45c5-9de1-d8bcba956088" (UID: "c9bd836e-9017-45c5-9de1-d8bcba956088"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.643643 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-utilities" (OuterVolumeSpecName: "utilities") pod "bbbbb850-78fa-47f0-8339-24148d9c610e" (UID: "bbbbb850-78fa-47f0-8339-24148d9c610e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.644350 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-utilities" (OuterVolumeSpecName: "utilities") pod "8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" (UID: "8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.646244 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e92734a4-fea9-42b6-badf-dc94c29db4fc-kube-api-access-lcbvz" (OuterVolumeSpecName: "kube-api-access-lcbvz") pod "e92734a4-fea9-42b6-badf-dc94c29db4fc" (UID: "e92734a4-fea9-42b6-badf-dc94c29db4fc"). InnerVolumeSpecName "kube-api-access-lcbvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.646505 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2d7a0ce6-fda6-4100-801b-a7170a4d5e93" (UID: "2d7a0ce6-fda6-4100-801b-a7170a4d5e93"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.646563 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbbbb850-78fa-47f0-8339-24148d9c610e-kube-api-access-9gtk9" (OuterVolumeSpecName: "kube-api-access-9gtk9") pod "bbbbb850-78fa-47f0-8339-24148d9c610e" (UID: "bbbbb850-78fa-47f0-8339-24148d9c610e"). InnerVolumeSpecName "kube-api-access-9gtk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.647038 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-utilities" (OuterVolumeSpecName: "utilities") pod "e92734a4-fea9-42b6-badf-dc94c29db4fc" (UID: "e92734a4-fea9-42b6-badf-dc94c29db4fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.648522 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-kube-api-access-ndfnk" (OuterVolumeSpecName: "kube-api-access-ndfnk") pod "8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" (UID: "8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9"). InnerVolumeSpecName "kube-api-access-ndfnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.648949 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-kube-api-access-865jv" (OuterVolumeSpecName: "kube-api-access-865jv") pod "2d7a0ce6-fda6-4100-801b-a7170a4d5e93" (UID: "2d7a0ce6-fda6-4100-801b-a7170a4d5e93"). InnerVolumeSpecName "kube-api-access-865jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.651937 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9bd836e-9017-45c5-9de1-d8bcba956088-kube-api-access-6lrb8" (OuterVolumeSpecName: "kube-api-access-6lrb8") pod "c9bd836e-9017-45c5-9de1-d8bcba956088" (UID: "c9bd836e-9017-45c5-9de1-d8bcba956088"). InnerVolumeSpecName "kube-api-access-6lrb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.676202 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" (UID: "8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.708354 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9bd836e-9017-45c5-9de1-d8bcba956088" (UID: "c9bd836e-9017-45c5-9de1-d8bcba956088"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.713382 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e92734a4-fea9-42b6-badf-dc94c29db4fc" (UID: "e92734a4-fea9-42b6-badf-dc94c29db4fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744251 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lrb8\" (UniqueName: \"kubernetes.io/projected/c9bd836e-9017-45c5-9de1-d8bcba956088-kube-api-access-6lrb8\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744278 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744287 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744300 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744309 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744317 4741 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744326 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcbvz\" (UniqueName: \"kubernetes.io/projected/e92734a4-fea9-42b6-badf-dc94c29db4fc-kube-api-access-lcbvz\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744334 4741 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744345 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744356 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndfnk\" (UniqueName: \"kubernetes.io/projected/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9-kube-api-access-ndfnk\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744363 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92734a4-fea9-42b6-badf-dc94c29db4fc-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744371 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-865jv\" (UniqueName: \"kubernetes.io/projected/2d7a0ce6-fda6-4100-801b-a7170a4d5e93-kube-api-access-865jv\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744380 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gtk9\" (UniqueName: \"kubernetes.io/projected/bbbbb850-78fa-47f0-8339-24148d9c610e-kube-api-access-9gtk9\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.744402 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9bd836e-9017-45c5-9de1-d8bcba956088-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.747121 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbbbb850-78fa-47f0-8339-24148d9c610e" (UID: "bbbbb850-78fa-47f0-8339-24148d9c610e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.845515 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbb850-78fa-47f0-8339-24148d9c610e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:13:43 crc kubenswrapper[4741]: I0929 19:13:43.957447 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fhs5q"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.185153 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c26bp" event={"ID":"c9bd836e-9017-45c5-9de1-d8bcba956088","Type":"ContainerDied","Data":"3ffe02a03ee97a11c66e710e5212465ad85046e55013133fa27f4b2e1151edc5"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.185198 4741 scope.go:117] "RemoveContainer" containerID="7116b971ecc2be02e630106b46ead40e546b84f444e211d4787d3b410bc58022" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.185195 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c26bp" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.188230 4741 generic.go:334] "Generic (PLEG): container finished" podID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerID="462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c" exitCode=0 Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.188293 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vmvz" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.188309 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vmvz" event={"ID":"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9","Type":"ContainerDied","Data":"462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.188355 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vmvz" event={"ID":"8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9","Type":"ContainerDied","Data":"779fb826d31d42bf2f2367020715fc7b009418ddce3be36e3f5955a410319add"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.191095 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kcsrb" event={"ID":"e92734a4-fea9-42b6-badf-dc94c29db4fc","Type":"ContainerDied","Data":"a82ef4d99c1fbd831d511122e1d38a67e148c51579f908ea888ab595f7d159ee"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.191126 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kcsrb" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.192858 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" event={"ID":"674ba05b-0612-46ca-b5b6-a91eb3950d27","Type":"ContainerStarted","Data":"081f0cc0be7ef795e088a7019b2af319cd08fbcde1cdf2f792971d0ebc64260b"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.192902 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" event={"ID":"674ba05b-0612-46ca-b5b6-a91eb3950d27","Type":"ContainerStarted","Data":"7a345e450477ccbaaa9ffdc854029dc1b0b991c7852dafd5cce705345ce83b43"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.193584 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.196147 4741 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fhs5q container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.196186 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" podUID="674ba05b-0612-46ca-b5b6-a91eb3950d27" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.196246 4741 generic.go:334] "Generic (PLEG): container finished" podID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerID="6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d" exitCode=0 Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.196312 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdvm7" event={"ID":"bbbbb850-78fa-47f0-8339-24148d9c610e","Type":"ContainerDied","Data":"6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.196338 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdvm7" event={"ID":"bbbbb850-78fa-47f0-8339-24148d9c610e","Type":"ContainerDied","Data":"15c7a9c498f0361f92aec09090766c78cf86c2e0c0e89e7065004c9d18575019"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.196456 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdvm7" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.202241 4741 generic.go:334] "Generic (PLEG): container finished" podID="2d7a0ce6-fda6-4100-801b-a7170a4d5e93" containerID="68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63" exitCode=0 Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.202279 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" event={"ID":"2d7a0ce6-fda6-4100-801b-a7170a4d5e93","Type":"ContainerDied","Data":"68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.202305 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" event={"ID":"2d7a0ce6-fda6-4100-801b-a7170a4d5e93","Type":"ContainerDied","Data":"64633855a131143b04a5b36e65a947c4bccfe8637320edd11453d0cf810e4911"} Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.202356 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fjfzz" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.225232 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" podStartSLOduration=1.225214098 podStartE2EDuration="1.225214098s" podCreationTimestamp="2025-09-29 19:13:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:13:44.217662137 +0000 UTC m=+265.865451479" watchObservedRunningTime="2025-09-29 19:13:44.225214098 +0000 UTC m=+265.873003430" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.254136 4741 scope.go:117] "RemoveContainer" containerID="723f531b35571bd73973e65853844683582ce74854dd892e051558ad0f171eac" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.279441 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fjfzz"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.283411 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fjfzz"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.289534 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vmvz"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.293660 4741 scope.go:117] "RemoveContainer" containerID="959f96147a6cb8490a090f8eae463193ac9fcc7203f4b8792eefa393fcff8f5f" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.296012 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vmvz"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.314541 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kcsrb"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.324339 4741 scope.go:117] "RemoveContainer" containerID="462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.324341 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kcsrb"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.329156 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c26bp"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.333761 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c26bp"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.340027 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdvm7"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.340289 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rdvm7"] Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.343127 4741 scope.go:117] "RemoveContainer" containerID="b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.358430 4741 scope.go:117] "RemoveContainer" containerID="3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.376840 4741 scope.go:117] "RemoveContainer" containerID="462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c" Sep 29 19:13:44 crc kubenswrapper[4741]: E0929 19:13:44.377262 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c\": container with ID starting with 462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c not found: ID does not exist" containerID="462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.377289 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c"} err="failed to get container status \"462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c\": rpc error: code = NotFound desc = could not find container \"462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c\": container with ID starting with 462bbdcea6354c1ec1888d4d3bb771bc85b5fca14cf085a36b5d964335e4ce3c not found: ID does not exist" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.377312 4741 scope.go:117] "RemoveContainer" containerID="b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d" Sep 29 19:13:44 crc kubenswrapper[4741]: E0929 19:13:44.377567 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d\": container with ID starting with b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d not found: ID does not exist" containerID="b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.377587 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d"} err="failed to get container status \"b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d\": rpc error: code = NotFound desc = could not find container \"b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d\": container with ID starting with b68e7508077493d4e7e025e919654de8c806cb494a5804b8dc312b6a7dee0a3d not found: ID does not exist" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.377601 4741 scope.go:117] "RemoveContainer" containerID="3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f" Sep 29 19:13:44 crc kubenswrapper[4741]: E0929 19:13:44.378036 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f\": container with ID starting with 3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f not found: ID does not exist" containerID="3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.378053 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f"} err="failed to get container status \"3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f\": rpc error: code = NotFound desc = could not find container \"3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f\": container with ID starting with 3ef7124a73c985dc8409bf1804abe7d5070bf0972a6ef2e9a3736bd4cbbcc21f not found: ID does not exist" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.378066 4741 scope.go:117] "RemoveContainer" containerID="0c36febf77e49538a39218224ce24f0835247300b12e21266ed524b41bb3b639" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.393804 4741 scope.go:117] "RemoveContainer" containerID="6591847a4bd829213f8292ce1506b707868e2e08f67cff424dd96b606ef349ca" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.409188 4741 scope.go:117] "RemoveContainer" containerID="110f431df0cca0b8961bdd50adf9023e2516a4935afe31b110b154779bf366a5" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.424014 4741 scope.go:117] "RemoveContainer" containerID="6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.438212 4741 scope.go:117] "RemoveContainer" containerID="9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.455689 4741 scope.go:117] "RemoveContainer" containerID="846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.475561 4741 scope.go:117] "RemoveContainer" containerID="6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d" Sep 29 19:13:44 crc kubenswrapper[4741]: E0929 19:13:44.476243 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d\": container with ID starting with 6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d not found: ID does not exist" containerID="6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.476312 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d"} err="failed to get container status \"6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d\": rpc error: code = NotFound desc = could not find container \"6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d\": container with ID starting with 6e092ef5367ca13ea5864638a16a96effb9ced56c131ab9580e1f5cf37a8d58d not found: ID does not exist" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.476360 4741 scope.go:117] "RemoveContainer" containerID="9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b" Sep 29 19:13:44 crc kubenswrapper[4741]: E0929 19:13:44.476928 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b\": container with ID starting with 9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b not found: ID does not exist" containerID="9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.476958 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b"} err="failed to get container status \"9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b\": rpc error: code = NotFound desc = could not find container \"9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b\": container with ID starting with 9db184a2998fad87f8bbe1fd6e9fccfe564c50f1076257ac38186f61791d719b not found: ID does not exist" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.476978 4741 scope.go:117] "RemoveContainer" containerID="846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376" Sep 29 19:13:44 crc kubenswrapper[4741]: E0929 19:13:44.477406 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376\": container with ID starting with 846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376 not found: ID does not exist" containerID="846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.477447 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376"} err="failed to get container status \"846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376\": rpc error: code = NotFound desc = could not find container \"846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376\": container with ID starting with 846713c256cd713cc8495a590f1601bdcf28bb0c03cd67ad98e2bc0b3736b376 not found: ID does not exist" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.477505 4741 scope.go:117] "RemoveContainer" containerID="68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.490097 4741 scope.go:117] "RemoveContainer" containerID="68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63" Sep 29 19:13:44 crc kubenswrapper[4741]: E0929 19:13:44.491530 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63\": container with ID starting with 68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63 not found: ID does not exist" containerID="68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63" Sep 29 19:13:44 crc kubenswrapper[4741]: I0929 19:13:44.491589 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63"} err="failed to get container status \"68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63\": rpc error: code = NotFound desc = could not find container \"68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63\": container with ID starting with 68e5a258a7fc28f085fab253212f1378dcfa126ab8ea77e89326535479182d63 not found: ID does not exist" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.092033 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d7a0ce6-fda6-4100-801b-a7170a4d5e93" path="/var/lib/kubelet/pods/2d7a0ce6-fda6-4100-801b-a7170a4d5e93/volumes" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.092619 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" path="/var/lib/kubelet/pods/8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9/volumes" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.093295 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" path="/var/lib/kubelet/pods/bbbbb850-78fa-47f0-8339-24148d9c610e/volumes" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.093994 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" path="/var/lib/kubelet/pods/c9bd836e-9017-45c5-9de1-d8bcba956088/volumes" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.094691 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" path="/var/lib/kubelet/pods/e92734a4-fea9-42b6-badf-dc94c29db4fc/volumes" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.221627 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fhs5q" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275521 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ztxkt"] Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275715 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275726 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275738 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275744 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275752 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275758 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275767 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275772 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275779 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275785 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275792 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d7a0ce6-fda6-4100-801b-a7170a4d5e93" containerName="marketplace-operator" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275799 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d7a0ce6-fda6-4100-801b-a7170a4d5e93" containerName="marketplace-operator" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275811 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275820 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275831 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275836 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275846 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275851 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="extract-content" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275859 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275864 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275876 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275881 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275889 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275895 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: E0929 19:13:45.275902 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275907 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="extract-utilities" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275981 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f69b21c-ce1f-4c27-8dcf-f71d2afcfee9" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.275992 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbbbb850-78fa-47f0-8339-24148d9c610e" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.276000 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9bd836e-9017-45c5-9de1-d8bcba956088" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.276007 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d7a0ce6-fda6-4100-801b-a7170a4d5e93" containerName="marketplace-operator" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.276013 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e92734a4-fea9-42b6-badf-dc94c29db4fc" containerName="registry-server" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.276658 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.283318 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.287808 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztxkt"] Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.365792 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df857807-c2a1-48da-9f68-ee9a9aafbb43-utilities\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.366327 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk5kw\" (UniqueName: \"kubernetes.io/projected/df857807-c2a1-48da-9f68-ee9a9aafbb43-kube-api-access-sk5kw\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.366463 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df857807-c2a1-48da-9f68-ee9a9aafbb43-catalog-content\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.467438 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bjmn7"] Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.468588 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.469604 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df857807-c2a1-48da-9f68-ee9a9aafbb43-catalog-content\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.469816 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df857807-c2a1-48da-9f68-ee9a9aafbb43-utilities\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.469852 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk5kw\" (UniqueName: \"kubernetes.io/projected/df857807-c2a1-48da-9f68-ee9a9aafbb43-kube-api-access-sk5kw\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.470576 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df857807-c2a1-48da-9f68-ee9a9aafbb43-catalog-content\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.470851 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df857807-c2a1-48da-9f68-ee9a9aafbb43-utilities\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.472977 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.481832 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bjmn7"] Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.491717 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk5kw\" (UniqueName: \"kubernetes.io/projected/df857807-c2a1-48da-9f68-ee9a9aafbb43-kube-api-access-sk5kw\") pod \"redhat-marketplace-ztxkt\" (UID: \"df857807-c2a1-48da-9f68-ee9a9aafbb43\") " pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.571623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c0cb0-3256-405d-a1ec-9884ef100eff-catalog-content\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.571669 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk6qd\" (UniqueName: \"kubernetes.io/projected/074c0cb0-3256-405d-a1ec-9884ef100eff-kube-api-access-xk6qd\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.571695 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c0cb0-3256-405d-a1ec-9884ef100eff-utilities\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.593543 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.677277 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c0cb0-3256-405d-a1ec-9884ef100eff-catalog-content\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.677320 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk6qd\" (UniqueName: \"kubernetes.io/projected/074c0cb0-3256-405d-a1ec-9884ef100eff-kube-api-access-xk6qd\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.677338 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c0cb0-3256-405d-a1ec-9884ef100eff-utilities\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.677813 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074c0cb0-3256-405d-a1ec-9884ef100eff-catalog-content\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.677909 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074c0cb0-3256-405d-a1ec-9884ef100eff-utilities\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.696142 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk6qd\" (UniqueName: \"kubernetes.io/projected/074c0cb0-3256-405d-a1ec-9884ef100eff-kube-api-access-xk6qd\") pod \"redhat-operators-bjmn7\" (UID: \"074c0cb0-3256-405d-a1ec-9884ef100eff\") " pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.786420 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:45 crc kubenswrapper[4741]: I0929 19:13:45.991084 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztxkt"] Sep 29 19:13:45 crc kubenswrapper[4741]: W0929 19:13:45.995484 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf857807_c2a1_48da_9f68_ee9a9aafbb43.slice/crio-f05c8a39f3507192fead82e5bffbd7deb2439aa5f3df5e32fc24c70b7b8bffbf WatchSource:0}: Error finding container f05c8a39f3507192fead82e5bffbd7deb2439aa5f3df5e32fc24c70b7b8bffbf: Status 404 returned error can't find the container with id f05c8a39f3507192fead82e5bffbd7deb2439aa5f3df5e32fc24c70b7b8bffbf Sep 29 19:13:46 crc kubenswrapper[4741]: I0929 19:13:46.177307 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bjmn7"] Sep 29 19:13:46 crc kubenswrapper[4741]: I0929 19:13:46.218831 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjmn7" event={"ID":"074c0cb0-3256-405d-a1ec-9884ef100eff","Type":"ContainerStarted","Data":"5b5208cc38c0966735b5e916f3de9b8237008075dde56be97d374c8ba0d50308"} Sep 29 19:13:46 crc kubenswrapper[4741]: I0929 19:13:46.224628 4741 generic.go:334] "Generic (PLEG): container finished" podID="df857807-c2a1-48da-9f68-ee9a9aafbb43" containerID="eaa9bfe2a5eb7b57ea48ccccfa0869a81f626fe116170c91a538efb2c5170847" exitCode=0 Sep 29 19:13:46 crc kubenswrapper[4741]: I0929 19:13:46.224730 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztxkt" event={"ID":"df857807-c2a1-48da-9f68-ee9a9aafbb43","Type":"ContainerDied","Data":"eaa9bfe2a5eb7b57ea48ccccfa0869a81f626fe116170c91a538efb2c5170847"} Sep 29 19:13:46 crc kubenswrapper[4741]: I0929 19:13:46.224847 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztxkt" event={"ID":"df857807-c2a1-48da-9f68-ee9a9aafbb43","Type":"ContainerStarted","Data":"f05c8a39f3507192fead82e5bffbd7deb2439aa5f3df5e32fc24c70b7b8bffbf"} Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.231260 4741 generic.go:334] "Generic (PLEG): container finished" podID="074c0cb0-3256-405d-a1ec-9884ef100eff" containerID="df54e314061ec44c24d922b792bc0bcee8cc8a5cb88f7538981452f7bfb9908f" exitCode=0 Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.231360 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjmn7" event={"ID":"074c0cb0-3256-405d-a1ec-9884ef100eff","Type":"ContainerDied","Data":"df54e314061ec44c24d922b792bc0bcee8cc8a5cb88f7538981452f7bfb9908f"} Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.673155 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k4kw7"] Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.674443 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.676748 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.687603 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4kw7"] Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.807231 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d135c17a-0ef7-4994-8125-019b0d6f4fde-catalog-content\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.807278 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d135c17a-0ef7-4994-8125-019b0d6f4fde-utilities\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.807369 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fjbg\" (UniqueName: \"kubernetes.io/projected/d135c17a-0ef7-4994-8125-019b0d6f4fde-kube-api-access-9fjbg\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.872990 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bkw22"] Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.873968 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.875748 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.883976 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkw22"] Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.908266 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d135c17a-0ef7-4994-8125-019b0d6f4fde-catalog-content\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.908310 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d135c17a-0ef7-4994-8125-019b0d6f4fde-utilities\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.908367 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fjbg\" (UniqueName: \"kubernetes.io/projected/d135c17a-0ef7-4994-8125-019b0d6f4fde-kube-api-access-9fjbg\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.909065 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d135c17a-0ef7-4994-8125-019b0d6f4fde-catalog-content\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.909080 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d135c17a-0ef7-4994-8125-019b0d6f4fde-utilities\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.926054 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fjbg\" (UniqueName: \"kubernetes.io/projected/d135c17a-0ef7-4994-8125-019b0d6f4fde-kube-api-access-9fjbg\") pod \"community-operators-k4kw7\" (UID: \"d135c17a-0ef7-4994-8125-019b0d6f4fde\") " pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:47 crc kubenswrapper[4741]: I0929 19:13:47.992354 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.009470 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/478c7368-6364-4c7f-89a3-9a6358c3318e-catalog-content\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.009541 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/478c7368-6364-4c7f-89a3-9a6358c3318e-utilities\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.009684 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xrdf\" (UniqueName: \"kubernetes.io/projected/478c7368-6364-4c7f-89a3-9a6358c3318e-kube-api-access-4xrdf\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.110973 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/478c7368-6364-4c7f-89a3-9a6358c3318e-catalog-content\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.111297 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/478c7368-6364-4c7f-89a3-9a6358c3318e-utilities\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.111333 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xrdf\" (UniqueName: \"kubernetes.io/projected/478c7368-6364-4c7f-89a3-9a6358c3318e-kube-api-access-4xrdf\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.111433 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/478c7368-6364-4c7f-89a3-9a6358c3318e-catalog-content\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.111759 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/478c7368-6364-4c7f-89a3-9a6358c3318e-utilities\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.126055 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xrdf\" (UniqueName: \"kubernetes.io/projected/478c7368-6364-4c7f-89a3-9a6358c3318e-kube-api-access-4xrdf\") pod \"certified-operators-bkw22\" (UID: \"478c7368-6364-4c7f-89a3-9a6358c3318e\") " pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.198932 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.242021 4741 generic.go:334] "Generic (PLEG): container finished" podID="df857807-c2a1-48da-9f68-ee9a9aafbb43" containerID="99eeec4d2f63c8ec1652a24417d210c245998d0c70b94b44f48b97c95326042f" exitCode=0 Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.242147 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztxkt" event={"ID":"df857807-c2a1-48da-9f68-ee9a9aafbb43","Type":"ContainerDied","Data":"99eeec4d2f63c8ec1652a24417d210c245998d0c70b94b44f48b97c95326042f"} Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.357002 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4kw7"] Sep 29 19:13:48 crc kubenswrapper[4741]: W0929 19:13:48.366523 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd135c17a_0ef7_4994_8125_019b0d6f4fde.slice/crio-3c59a733bd774c1ba3abfe6b4834b2c5da0b57fa5ae9839d8256079749e918da WatchSource:0}: Error finding container 3c59a733bd774c1ba3abfe6b4834b2c5da0b57fa5ae9839d8256079749e918da: Status 404 returned error can't find the container with id 3c59a733bd774c1ba3abfe6b4834b2c5da0b57fa5ae9839d8256079749e918da Sep 29 19:13:48 crc kubenswrapper[4741]: I0929 19:13:48.590011 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkw22"] Sep 29 19:13:48 crc kubenswrapper[4741]: W0929 19:13:48.598420 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod478c7368_6364_4c7f_89a3_9a6358c3318e.slice/crio-fe644f41ddd779ad3d6c27043d85a610ebe54c75d95ac9f0cafaab2b781da006 WatchSource:0}: Error finding container fe644f41ddd779ad3d6c27043d85a610ebe54c75d95ac9f0cafaab2b781da006: Status 404 returned error can't find the container with id fe644f41ddd779ad3d6c27043d85a610ebe54c75d95ac9f0cafaab2b781da006 Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.248381 4741 generic.go:334] "Generic (PLEG): container finished" podID="d135c17a-0ef7-4994-8125-019b0d6f4fde" containerID="dc35555ec06350e74306b4bed3f3247a731d6f1ed524dacc5172be699eb9df76" exitCode=0 Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.248600 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4kw7" event={"ID":"d135c17a-0ef7-4994-8125-019b0d6f4fde","Type":"ContainerDied","Data":"dc35555ec06350e74306b4bed3f3247a731d6f1ed524dacc5172be699eb9df76"} Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.248800 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4kw7" event={"ID":"d135c17a-0ef7-4994-8125-019b0d6f4fde","Type":"ContainerStarted","Data":"3c59a733bd774c1ba3abfe6b4834b2c5da0b57fa5ae9839d8256079749e918da"} Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.255174 4741 generic.go:334] "Generic (PLEG): container finished" podID="074c0cb0-3256-405d-a1ec-9884ef100eff" containerID="70d865147f6b591a9919d21a4ea83257c2e85fe823cf56abb91e964f00f7d970" exitCode=0 Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.255232 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjmn7" event={"ID":"074c0cb0-3256-405d-a1ec-9884ef100eff","Type":"ContainerDied","Data":"70d865147f6b591a9919d21a4ea83257c2e85fe823cf56abb91e964f00f7d970"} Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.259376 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztxkt" event={"ID":"df857807-c2a1-48da-9f68-ee9a9aafbb43","Type":"ContainerStarted","Data":"c983e1c86bd976e4bcae9421c504f79e91f0f3c84bb5afa32281744b0e8ddd47"} Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.261993 4741 generic.go:334] "Generic (PLEG): container finished" podID="478c7368-6364-4c7f-89a3-9a6358c3318e" containerID="b7b9400c0cc84ed4ead97bf3eb6d55cb074c53c6415341c6c0d1e8f8237b907e" exitCode=0 Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.262047 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkw22" event={"ID":"478c7368-6364-4c7f-89a3-9a6358c3318e","Type":"ContainerDied","Data":"b7b9400c0cc84ed4ead97bf3eb6d55cb074c53c6415341c6c0d1e8f8237b907e"} Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.262075 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkw22" event={"ID":"478c7368-6364-4c7f-89a3-9a6358c3318e","Type":"ContainerStarted","Data":"fe644f41ddd779ad3d6c27043d85a610ebe54c75d95ac9f0cafaab2b781da006"} Sep 29 19:13:49 crc kubenswrapper[4741]: I0929 19:13:49.339590 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ztxkt" podStartSLOduration=1.88531692 podStartE2EDuration="4.339569515s" podCreationTimestamp="2025-09-29 19:13:45 +0000 UTC" firstStartedPulling="2025-09-29 19:13:46.225935424 +0000 UTC m=+267.873724756" lastFinishedPulling="2025-09-29 19:13:48.680188019 +0000 UTC m=+270.327977351" observedRunningTime="2025-09-29 19:13:49.337940223 +0000 UTC m=+270.985729565" watchObservedRunningTime="2025-09-29 19:13:49.339569515 +0000 UTC m=+270.987358847" Sep 29 19:13:50 crc kubenswrapper[4741]: I0929 19:13:50.268137 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4kw7" event={"ID":"d135c17a-0ef7-4994-8125-019b0d6f4fde","Type":"ContainerStarted","Data":"764d47d4a089746c3c4c709582b3811801db6e3ae4d15d3666b64d0edb4fff37"} Sep 29 19:13:50 crc kubenswrapper[4741]: I0929 19:13:50.277271 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjmn7" event={"ID":"074c0cb0-3256-405d-a1ec-9884ef100eff","Type":"ContainerStarted","Data":"69fbda2f85880a01d842c9a0f7e4a1c31baafe864f2949830d8e390acefda96b"} Sep 29 19:13:50 crc kubenswrapper[4741]: I0929 19:13:50.279671 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkw22" event={"ID":"478c7368-6364-4c7f-89a3-9a6358c3318e","Type":"ContainerStarted","Data":"b4cbe3eb824ef053e03d91378fa739db42c15b67317d6d92bcb80307b1f49f1d"} Sep 29 19:13:50 crc kubenswrapper[4741]: I0929 19:13:50.329571 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bjmn7" podStartSLOduration=2.882900556 podStartE2EDuration="5.329552738s" podCreationTimestamp="2025-09-29 19:13:45 +0000 UTC" firstStartedPulling="2025-09-29 19:13:47.232559597 +0000 UTC m=+268.880348929" lastFinishedPulling="2025-09-29 19:13:49.679211779 +0000 UTC m=+271.327001111" observedRunningTime="2025-09-29 19:13:50.328899487 +0000 UTC m=+271.976688829" watchObservedRunningTime="2025-09-29 19:13:50.329552738 +0000 UTC m=+271.977342070" Sep 29 19:13:51 crc kubenswrapper[4741]: I0929 19:13:51.285652 4741 generic.go:334] "Generic (PLEG): container finished" podID="d135c17a-0ef7-4994-8125-019b0d6f4fde" containerID="764d47d4a089746c3c4c709582b3811801db6e3ae4d15d3666b64d0edb4fff37" exitCode=0 Sep 29 19:13:51 crc kubenswrapper[4741]: I0929 19:13:51.286012 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4kw7" event={"ID":"d135c17a-0ef7-4994-8125-019b0d6f4fde","Type":"ContainerDied","Data":"764d47d4a089746c3c4c709582b3811801db6e3ae4d15d3666b64d0edb4fff37"} Sep 29 19:13:51 crc kubenswrapper[4741]: I0929 19:13:51.292033 4741 generic.go:334] "Generic (PLEG): container finished" podID="478c7368-6364-4c7f-89a3-9a6358c3318e" containerID="b4cbe3eb824ef053e03d91378fa739db42c15b67317d6d92bcb80307b1f49f1d" exitCode=0 Sep 29 19:13:51 crc kubenswrapper[4741]: I0929 19:13:51.292083 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkw22" event={"ID":"478c7368-6364-4c7f-89a3-9a6358c3318e","Type":"ContainerDied","Data":"b4cbe3eb824ef053e03d91378fa739db42c15b67317d6d92bcb80307b1f49f1d"} Sep 29 19:13:52 crc kubenswrapper[4741]: I0929 19:13:52.299555 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4kw7" event={"ID":"d135c17a-0ef7-4994-8125-019b0d6f4fde","Type":"ContainerStarted","Data":"4c8a9575cc8185375ecc36f5e725b781560178821686520cbea1cd8389d64764"} Sep 29 19:13:52 crc kubenswrapper[4741]: I0929 19:13:52.301754 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkw22" event={"ID":"478c7368-6364-4c7f-89a3-9a6358c3318e","Type":"ContainerStarted","Data":"9ad9f941c96b905a9ea2acb4f20d19849591026e8fdbb55f8b8a113c82065f30"} Sep 29 19:13:52 crc kubenswrapper[4741]: I0929 19:13:52.317766 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k4kw7" podStartSLOduration=2.873502471 podStartE2EDuration="5.317746767s" podCreationTimestamp="2025-09-29 19:13:47 +0000 UTC" firstStartedPulling="2025-09-29 19:13:49.25043301 +0000 UTC m=+270.898222342" lastFinishedPulling="2025-09-29 19:13:51.694677306 +0000 UTC m=+273.342466638" observedRunningTime="2025-09-29 19:13:52.317118437 +0000 UTC m=+273.964907779" watchObservedRunningTime="2025-09-29 19:13:52.317746767 +0000 UTC m=+273.965536099" Sep 29 19:13:52 crc kubenswrapper[4741]: I0929 19:13:52.333834 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bkw22" podStartSLOduration=2.917295133 podStartE2EDuration="5.333813098s" podCreationTimestamp="2025-09-29 19:13:47 +0000 UTC" firstStartedPulling="2025-09-29 19:13:49.263442423 +0000 UTC m=+270.911231755" lastFinishedPulling="2025-09-29 19:13:51.679960368 +0000 UTC m=+273.327749720" observedRunningTime="2025-09-29 19:13:52.332807946 +0000 UTC m=+273.980597268" watchObservedRunningTime="2025-09-29 19:13:52.333813098 +0000 UTC m=+273.981602430" Sep 29 19:13:55 crc kubenswrapper[4741]: I0929 19:13:55.594083 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:55 crc kubenswrapper[4741]: I0929 19:13:55.594730 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:55 crc kubenswrapper[4741]: I0929 19:13:55.635721 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:55 crc kubenswrapper[4741]: I0929 19:13:55.788056 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:55 crc kubenswrapper[4741]: I0929 19:13:55.788273 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:55 crc kubenswrapper[4741]: I0929 19:13:55.847718 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:56 crc kubenswrapper[4741]: I0929 19:13:56.364861 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ztxkt" Sep 29 19:13:56 crc kubenswrapper[4741]: I0929 19:13:56.365313 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bjmn7" Sep 29 19:13:57 crc kubenswrapper[4741]: I0929 19:13:57.992774 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:57 crc kubenswrapper[4741]: I0929 19:13:57.992841 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:58 crc kubenswrapper[4741]: I0929 19:13:58.034927 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:58 crc kubenswrapper[4741]: I0929 19:13:58.199346 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:58 crc kubenswrapper[4741]: I0929 19:13:58.199407 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:58 crc kubenswrapper[4741]: I0929 19:13:58.238292 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:13:58 crc kubenswrapper[4741]: I0929 19:13:58.368438 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k4kw7" Sep 29 19:13:58 crc kubenswrapper[4741]: I0929 19:13:58.372013 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bkw22" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.137246 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d"] Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.138861 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.141925 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.142308 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.155346 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d"] Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.247507 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-secret-volume\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.247580 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7fml\" (UniqueName: \"kubernetes.io/projected/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-kube-api-access-v7fml\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.247623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-config-volume\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.348734 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-secret-volume\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.348800 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7fml\" (UniqueName: \"kubernetes.io/projected/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-kube-api-access-v7fml\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.348840 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-config-volume\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.350347 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-config-volume\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.354132 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-secret-volume\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.373369 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7fml\" (UniqueName: \"kubernetes.io/projected/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-kube-api-access-v7fml\") pod \"collect-profiles-29319555-knl4d\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.462093 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:00 crc kubenswrapper[4741]: I0929 19:15:00.670048 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d"] Sep 29 19:15:01 crc kubenswrapper[4741]: I0929 19:15:01.680706 4741 generic.go:334] "Generic (PLEG): container finished" podID="e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" containerID="ac53733f11090f6bc9ed473e41f09c141c1505335e4106baba8391c443a7c356" exitCode=0 Sep 29 19:15:01 crc kubenswrapper[4741]: I0929 19:15:01.680773 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" event={"ID":"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e","Type":"ContainerDied","Data":"ac53733f11090f6bc9ed473e41f09c141c1505335e4106baba8391c443a7c356"} Sep 29 19:15:01 crc kubenswrapper[4741]: I0929 19:15:01.681012 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" event={"ID":"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e","Type":"ContainerStarted","Data":"de4f14265a0e16e53d5ad5d62bb94ecacb4e0e313d976b7ee6ef731b9bf8df0c"} Sep 29 19:15:02 crc kubenswrapper[4741]: I0929 19:15:02.905029 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.105548 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7fml\" (UniqueName: \"kubernetes.io/projected/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-kube-api-access-v7fml\") pod \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.105611 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-secret-volume\") pod \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.105651 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-config-volume\") pod \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\" (UID: \"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e\") " Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.106626 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-config-volume" (OuterVolumeSpecName: "config-volume") pod "e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" (UID: "e0ff83e8-1d40-4ae2-9878-24ed37e22c3e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.110104 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" (UID: "e0ff83e8-1d40-4ae2-9878-24ed37e22c3e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.110236 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-kube-api-access-v7fml" (OuterVolumeSpecName: "kube-api-access-v7fml") pod "e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" (UID: "e0ff83e8-1d40-4ae2-9878-24ed37e22c3e"). InnerVolumeSpecName "kube-api-access-v7fml". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.206676 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7fml\" (UniqueName: \"kubernetes.io/projected/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-kube-api-access-v7fml\") on node \"crc\" DevicePath \"\"" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.206712 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.206724 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.690968 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" event={"ID":"e0ff83e8-1d40-4ae2-9878-24ed37e22c3e","Type":"ContainerDied","Data":"de4f14265a0e16e53d5ad5d62bb94ecacb4e0e313d976b7ee6ef731b9bf8df0c"} Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.691015 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de4f14265a0e16e53d5ad5d62bb94ecacb4e0e313d976b7ee6ef731b9bf8df0c" Sep 29 19:15:03 crc kubenswrapper[4741]: I0929 19:15:03.691102 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d" Sep 29 19:15:31 crc kubenswrapper[4741]: I0929 19:15:31.739198 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:15:31 crc kubenswrapper[4741]: I0929 19:15:31.739679 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:16:01 crc kubenswrapper[4741]: I0929 19:16:01.739031 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:16:01 crc kubenswrapper[4741]: I0929 19:16:01.739566 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:16:31 crc kubenswrapper[4741]: I0929 19:16:31.739714 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:16:31 crc kubenswrapper[4741]: I0929 19:16:31.740475 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:16:31 crc kubenswrapper[4741]: I0929 19:16:31.740532 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:16:31 crc kubenswrapper[4741]: I0929 19:16:31.741277 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d272b5f97846d5ba906fda7647c908fb192dc3a478a94f6b285f09880ecc623"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:16:31 crc kubenswrapper[4741]: I0929 19:16:31.741333 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://7d272b5f97846d5ba906fda7647c908fb192dc3a478a94f6b285f09880ecc623" gracePeriod=600 Sep 29 19:16:32 crc kubenswrapper[4741]: I0929 19:16:32.171972 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="7d272b5f97846d5ba906fda7647c908fb192dc3a478a94f6b285f09880ecc623" exitCode=0 Sep 29 19:16:32 crc kubenswrapper[4741]: I0929 19:16:32.172042 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"7d272b5f97846d5ba906fda7647c908fb192dc3a478a94f6b285f09880ecc623"} Sep 29 19:16:32 crc kubenswrapper[4741]: I0929 19:16:32.172567 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"fa3477edf9ba3614f07b8cc994bf0edeaf4d0348d0bf19dc3fa71df97308d735"} Sep 29 19:16:32 crc kubenswrapper[4741]: I0929 19:16:32.172609 4741 scope.go:117] "RemoveContainer" containerID="946d11fb1bacd0e3828bbc6174414428118402a9c31273e1a99b26f8581ce2e2" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.815273 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-78dhg"] Sep 29 19:16:38 crc kubenswrapper[4741]: E0929 19:16:38.816151 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" containerName="collect-profiles" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.816169 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" containerName="collect-profiles" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.816274 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" containerName="collect-profiles" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.816711 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.834625 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-78dhg"] Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960102 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-bound-sa-token\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960173 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a906fb02-b396-4354-95e4-6de2f404c0e8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960209 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a906fb02-b396-4354-95e4-6de2f404c0e8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960231 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-kube-api-access-ljqsg\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960253 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a906fb02-b396-4354-95e4-6de2f404c0e8-trusted-ca\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960278 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-registry-tls\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960307 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a906fb02-b396-4354-95e4-6de2f404c0e8-registry-certificates\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.960351 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:38 crc kubenswrapper[4741]: I0929 19:16:38.979423 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.061849 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-bound-sa-token\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.061904 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a906fb02-b396-4354-95e4-6de2f404c0e8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.061936 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a906fb02-b396-4354-95e4-6de2f404c0e8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.061959 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-kube-api-access-ljqsg\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.061985 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a906fb02-b396-4354-95e4-6de2f404c0e8-trusted-ca\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.062032 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-registry-tls\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.062057 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a906fb02-b396-4354-95e4-6de2f404c0e8-registry-certificates\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.063744 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a906fb02-b396-4354-95e4-6de2f404c0e8-registry-certificates\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.065444 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a906fb02-b396-4354-95e4-6de2f404c0e8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.068441 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a906fb02-b396-4354-95e4-6de2f404c0e8-trusted-ca\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.071676 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a906fb02-b396-4354-95e4-6de2f404c0e8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.071813 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-registry-tls\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.078381 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-bound-sa-token\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.081153 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/a906fb02-b396-4354-95e4-6de2f404c0e8-kube-api-access-ljqsg\") pod \"image-registry-66df7c8f76-78dhg\" (UID: \"a906fb02-b396-4354-95e4-6de2f404c0e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.132718 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:39 crc kubenswrapper[4741]: I0929 19:16:39.341508 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-78dhg"] Sep 29 19:16:39 crc kubenswrapper[4741]: W0929 19:16:39.355852 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda906fb02_b396_4354_95e4_6de2f404c0e8.slice/crio-7fb84d67dd3ab2dc317ab416ecfdfe2bf02ff6e5e6f8fb3d2f84ccbcf777d96c WatchSource:0}: Error finding container 7fb84d67dd3ab2dc317ab416ecfdfe2bf02ff6e5e6f8fb3d2f84ccbcf777d96c: Status 404 returned error can't find the container with id 7fb84d67dd3ab2dc317ab416ecfdfe2bf02ff6e5e6f8fb3d2f84ccbcf777d96c Sep 29 19:16:40 crc kubenswrapper[4741]: I0929 19:16:40.217678 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" event={"ID":"a906fb02-b396-4354-95e4-6de2f404c0e8","Type":"ContainerStarted","Data":"05f09e5084f90e3bc264c4026c5a15d91e19def7cd2d22ad21290ed97fcb3dcf"} Sep 29 19:16:40 crc kubenswrapper[4741]: I0929 19:16:40.218081 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:40 crc kubenswrapper[4741]: I0929 19:16:40.218094 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" event={"ID":"a906fb02-b396-4354-95e4-6de2f404c0e8","Type":"ContainerStarted","Data":"7fb84d67dd3ab2dc317ab416ecfdfe2bf02ff6e5e6f8fb3d2f84ccbcf777d96c"} Sep 29 19:16:40 crc kubenswrapper[4741]: I0929 19:16:40.242031 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" podStartSLOduration=2.241980906 podStartE2EDuration="2.241980906s" podCreationTimestamp="2025-09-29 19:16:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:16:40.237824075 +0000 UTC m=+441.885613427" watchObservedRunningTime="2025-09-29 19:16:40.241980906 +0000 UTC m=+441.889770258" Sep 29 19:16:59 crc kubenswrapper[4741]: I0929 19:16:59.139698 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-78dhg" Sep 29 19:16:59 crc kubenswrapper[4741]: I0929 19:16:59.192171 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-d4ltl"] Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.246507 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" podUID="f05d1ab9-cf4d-4a05-a4bd-2d029564c266" containerName="registry" containerID="cri-o://6c4baea219594446fc8f4e051043f3678f8e20a2a368cd2cbbeaf711c2804e6f" gracePeriod=30 Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.485147 4741 generic.go:334] "Generic (PLEG): container finished" podID="f05d1ab9-cf4d-4a05-a4bd-2d029564c266" containerID="6c4baea219594446fc8f4e051043f3678f8e20a2a368cd2cbbeaf711c2804e6f" exitCode=0 Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.485240 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" event={"ID":"f05d1ab9-cf4d-4a05-a4bd-2d029564c266","Type":"ContainerDied","Data":"6c4baea219594446fc8f4e051043f3678f8e20a2a368cd2cbbeaf711c2804e6f"} Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.566474 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.702588 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-trusted-ca\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.702679 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-ca-trust-extracted\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.702712 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5q8m\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-kube-api-access-p5q8m\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.702959 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.703015 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-bound-sa-token\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.703045 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-installation-pull-secrets\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.703101 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-tls\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.703141 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-certificates\") pod \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\" (UID: \"f05d1ab9-cf4d-4a05-a4bd-2d029564c266\") " Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.704031 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.704196 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.708452 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-kube-api-access-p5q8m" (OuterVolumeSpecName: "kube-api-access-p5q8m") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "kube-api-access-p5q8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.711133 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.711600 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.712198 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.712635 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.720376 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "f05d1ab9-cf4d-4a05-a4bd-2d029564c266" (UID: "f05d1ab9-cf4d-4a05-a4bd-2d029564c266"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.804266 4741 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.804305 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.804317 4741 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.804328 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5q8m\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-kube-api-access-p5q8m\") on node \"crc\" DevicePath \"\"" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.804336 4741 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.804347 4741 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 19:17:24 crc kubenswrapper[4741]: I0929 19:17:24.804354 4741 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f05d1ab9-cf4d-4a05-a4bd-2d029564c266-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:17:25 crc kubenswrapper[4741]: I0929 19:17:25.501638 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" event={"ID":"f05d1ab9-cf4d-4a05-a4bd-2d029564c266","Type":"ContainerDied","Data":"9780b9423bf6cdd730f796fdcc223cf3d4fa7b87fb60b316a7624fbd36c05e63"} Sep 29 19:17:25 crc kubenswrapper[4741]: I0929 19:17:25.501702 4741 scope.go:117] "RemoveContainer" containerID="6c4baea219594446fc8f4e051043f3678f8e20a2a368cd2cbbeaf711c2804e6f" Sep 29 19:17:25 crc kubenswrapper[4741]: I0929 19:17:25.501700 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-d4ltl" Sep 29 19:17:25 crc kubenswrapper[4741]: I0929 19:17:25.518151 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-d4ltl"] Sep 29 19:17:25 crc kubenswrapper[4741]: I0929 19:17:25.521805 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-d4ltl"] Sep 29 19:17:27 crc kubenswrapper[4741]: I0929 19:17:27.094593 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f05d1ab9-cf4d-4a05-a4bd-2d029564c266" path="/var/lib/kubelet/pods/f05d1ab9-cf4d-4a05-a4bd-2d029564c266/volumes" Sep 29 19:18:31 crc kubenswrapper[4741]: I0929 19:18:31.738829 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:18:31 crc kubenswrapper[4741]: I0929 19:18:31.739553 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:19:01 crc kubenswrapper[4741]: I0929 19:19:01.739038 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:19:01 crc kubenswrapper[4741]: I0929 19:19:01.740687 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:19:31 crc kubenswrapper[4741]: I0929 19:19:31.738805 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:19:31 crc kubenswrapper[4741]: I0929 19:19:31.739410 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:19:31 crc kubenswrapper[4741]: I0929 19:19:31.739460 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:19:31 crc kubenswrapper[4741]: I0929 19:19:31.740104 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa3477edf9ba3614f07b8cc994bf0edeaf4d0348d0bf19dc3fa71df97308d735"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:19:31 crc kubenswrapper[4741]: I0929 19:19:31.740167 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://fa3477edf9ba3614f07b8cc994bf0edeaf4d0348d0bf19dc3fa71df97308d735" gracePeriod=600 Sep 29 19:19:32 crc kubenswrapper[4741]: I0929 19:19:32.188551 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="fa3477edf9ba3614f07b8cc994bf0edeaf4d0348d0bf19dc3fa71df97308d735" exitCode=0 Sep 29 19:19:32 crc kubenswrapper[4741]: I0929 19:19:32.189135 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"fa3477edf9ba3614f07b8cc994bf0edeaf4d0348d0bf19dc3fa71df97308d735"} Sep 29 19:19:32 crc kubenswrapper[4741]: I0929 19:19:32.189163 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"1f264b13b6ab5a53b78801880659a4525cf694d3e6348258f41be568cdac61c6"} Sep 29 19:19:32 crc kubenswrapper[4741]: I0929 19:19:32.189179 4741 scope.go:117] "RemoveContainer" containerID="7d272b5f97846d5ba906fda7647c908fb192dc3a478a94f6b285f09880ecc623" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.482887 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-6mdn7"] Sep 29 19:20:35 crc kubenswrapper[4741]: E0929 19:20:35.484086 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05d1ab9-cf4d-4a05-a4bd-2d029564c266" containerName="registry" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.484111 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05d1ab9-cf4d-4a05-a4bd-2d029564c266" containerName="registry" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.484306 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f05d1ab9-cf4d-4a05-a4bd-2d029564c266" containerName="registry" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.484898 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.488025 4741 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-q2245" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.488317 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.488500 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.488516 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.490658 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-6mdn7"] Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.608489 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f516a891-abe1-4143-9492-7bd23d5e74a9-node-mnt\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.608566 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f516a891-abe1-4143-9492-7bd23d5e74a9-crc-storage\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.608618 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47rb6\" (UniqueName: \"kubernetes.io/projected/f516a891-abe1-4143-9492-7bd23d5e74a9-kube-api-access-47rb6\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.710015 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47rb6\" (UniqueName: \"kubernetes.io/projected/f516a891-abe1-4143-9492-7bd23d5e74a9-kube-api-access-47rb6\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.710100 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f516a891-abe1-4143-9492-7bd23d5e74a9-node-mnt\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.710135 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f516a891-abe1-4143-9492-7bd23d5e74a9-crc-storage\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.710537 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f516a891-abe1-4143-9492-7bd23d5e74a9-node-mnt\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.710908 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f516a891-abe1-4143-9492-7bd23d5e74a9-crc-storage\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.731309 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47rb6\" (UniqueName: \"kubernetes.io/projected/f516a891-abe1-4143-9492-7bd23d5e74a9-kube-api-access-47rb6\") pod \"crc-storage-crc-6mdn7\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.812035 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.983558 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-6mdn7"] Sep 29 19:20:35 crc kubenswrapper[4741]: I0929 19:20:35.991074 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 19:20:36 crc kubenswrapper[4741]: I0929 19:20:36.582649 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-6mdn7" event={"ID":"f516a891-abe1-4143-9492-7bd23d5e74a9","Type":"ContainerStarted","Data":"76aa2c4aa184a8d99c54f7a06c764d9cb7ce64943c0862f18ae4b44d4909fb0d"} Sep 29 19:20:37 crc kubenswrapper[4741]: I0929 19:20:37.590545 4741 generic.go:334] "Generic (PLEG): container finished" podID="f516a891-abe1-4143-9492-7bd23d5e74a9" containerID="bc08b41c9620475f74ab11f5fe156357925b0bb34e7c96e5b73edf6a30202be6" exitCode=0 Sep 29 19:20:37 crc kubenswrapper[4741]: I0929 19:20:37.590608 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-6mdn7" event={"ID":"f516a891-abe1-4143-9492-7bd23d5e74a9","Type":"ContainerDied","Data":"bc08b41c9620475f74ab11f5fe156357925b0bb34e7c96e5b73edf6a30202be6"} Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.801486 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.947821 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f516a891-abe1-4143-9492-7bd23d5e74a9-crc-storage\") pod \"f516a891-abe1-4143-9492-7bd23d5e74a9\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.947905 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f516a891-abe1-4143-9492-7bd23d5e74a9-node-mnt\") pod \"f516a891-abe1-4143-9492-7bd23d5e74a9\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.947954 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47rb6\" (UniqueName: \"kubernetes.io/projected/f516a891-abe1-4143-9492-7bd23d5e74a9-kube-api-access-47rb6\") pod \"f516a891-abe1-4143-9492-7bd23d5e74a9\" (UID: \"f516a891-abe1-4143-9492-7bd23d5e74a9\") " Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.948097 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f516a891-abe1-4143-9492-7bd23d5e74a9-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "f516a891-abe1-4143-9492-7bd23d5e74a9" (UID: "f516a891-abe1-4143-9492-7bd23d5e74a9"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.948230 4741 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f516a891-abe1-4143-9492-7bd23d5e74a9-node-mnt\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.953309 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f516a891-abe1-4143-9492-7bd23d5e74a9-kube-api-access-47rb6" (OuterVolumeSpecName: "kube-api-access-47rb6") pod "f516a891-abe1-4143-9492-7bd23d5e74a9" (UID: "f516a891-abe1-4143-9492-7bd23d5e74a9"). InnerVolumeSpecName "kube-api-access-47rb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:20:38 crc kubenswrapper[4741]: I0929 19:20:38.961901 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f516a891-abe1-4143-9492-7bd23d5e74a9-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "f516a891-abe1-4143-9492-7bd23d5e74a9" (UID: "f516a891-abe1-4143-9492-7bd23d5e74a9"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.049907 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47rb6\" (UniqueName: \"kubernetes.io/projected/f516a891-abe1-4143-9492-7bd23d5e74a9-kube-api-access-47rb6\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.049954 4741 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f516a891-abe1-4143-9492-7bd23d5e74a9-crc-storage\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.273352 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mrdpd"] Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.273849 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-controller" containerID="cri-o://d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.273888 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="nbdb" containerID="cri-o://3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.273977 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="northd" containerID="cri-o://19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.274033 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.274032 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="sbdb" containerID="cri-o://0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.274103 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-acl-logging" containerID="cri-o://8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.274155 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-node" containerID="cri-o://68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.313364 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" containerID="cri-o://3637cdd09ff11428d033a392b90d98862680635742030c3645664937ffd7b6b2" gracePeriod=30 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.603337 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovnkube-controller/3.log" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.606204 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovn-acl-logging/0.log" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.606693 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovn-controller/0.log" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607083 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="3637cdd09ff11428d033a392b90d98862680635742030c3645664937ffd7b6b2" exitCode=0 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607113 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869" exitCode=0 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607123 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be" exitCode=0 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607133 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61" exitCode=0 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607142 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814" exitCode=0 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607151 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1" exitCode=0 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607149 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"3637cdd09ff11428d033a392b90d98862680635742030c3645664937ffd7b6b2"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607209 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607220 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607229 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607238 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607247 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607256 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607272 4741 scope.go:117] "RemoveContainer" containerID="db25cc37238c193f09967a1bef8528f05a3bd4f25991f83a3826ad40ff6446cd" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607159 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520" exitCode=143 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607341 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerID="d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff" exitCode=143 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.607355 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.610444 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6mdn7" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.610487 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-6mdn7" event={"ID":"f516a891-abe1-4143-9492-7bd23d5e74a9","Type":"ContainerDied","Data":"76aa2c4aa184a8d99c54f7a06c764d9cb7ce64943c0862f18ae4b44d4909fb0d"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.610506 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76aa2c4aa184a8d99c54f7a06c764d9cb7ce64943c0862f18ae4b44d4909fb0d" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.612358 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/2.log" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.612962 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/1.log" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.613005 4741 generic.go:334] "Generic (PLEG): container finished" podID="565a367b-9a7a-4b93-b6f8-638b639fef9b" containerID="472e7bb64bf7085f0ba716a2fc6cc1eb722f462abeeec7bbfddd1fc9cfcc3941" exitCode=2 Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.613061 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerDied","Data":"472e7bb64bf7085f0ba716a2fc6cc1eb722f462abeeec7bbfddd1fc9cfcc3941"} Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.613684 4741 scope.go:117] "RemoveContainer" containerID="472e7bb64bf7085f0ba716a2fc6cc1eb722f462abeeec7bbfddd1fc9cfcc3941" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.613894 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-ql4kz_openshift-multus(565a367b-9a7a-4b93-b6f8-638b639fef9b)\"" pod="openshift-multus/multus-ql4kz" podUID="565a367b-9a7a-4b93-b6f8-638b639fef9b" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.631665 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovn-acl-logging/0.log" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.632311 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovn-controller/0.log" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.632845 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.643870 4741 scope.go:117] "RemoveContainer" containerID="cc23359d0d3a3a290c0b5dac57441eaf7077826681af81be70095275dc1729f1" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.684871 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c4ps4"] Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685066 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="sbdb" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685082 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="sbdb" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685090 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685098 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685106 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685114 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685120 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685125 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685133 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-acl-logging" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685139 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-acl-logging" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685146 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="nbdb" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685152 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="nbdb" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685159 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f516a891-abe1-4143-9492-7bd23d5e74a9" containerName="storage" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685165 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f516a891-abe1-4143-9492-7bd23d5e74a9" containerName="storage" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685175 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685181 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685188 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-node" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685195 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-node" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685203 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kubecfg-setup" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685209 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kubecfg-setup" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685219 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685225 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685232 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685237 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685250 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="northd" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685256 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="northd" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685343 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685353 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685365 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f516a891-abe1-4143-9492-7bd23d5e74a9" containerName="storage" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685372 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685382 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685410 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovn-acl-logging" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685420 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685428 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="nbdb" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685436 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="kube-rbac-proxy-node" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685445 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="northd" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685453 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="sbdb" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685462 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: E0929 19:20:39.685563 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685572 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.685671 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" containerName="ovnkube-controller" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.687180 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758362 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-bin\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758424 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-systemd-units\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758486 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-netd\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758510 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-systemd\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758531 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-kubelet\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758566 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhpnc\" (UniqueName: \"kubernetes.io/projected/7a31e12c-13a1-4052-b679-7115e1a3574e-kube-api-access-fhpnc\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758605 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a31e12c-13a1-4052-b679-7115e1a3574e-ovn-node-metrics-cert\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758634 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758656 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-var-lib-openvswitch\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758683 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-config\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758705 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-etc-openvswitch\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758737 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-env-overrides\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758759 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-log-socket\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758780 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-ovn\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758799 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-node-log\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758872 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-slash\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758928 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-ovn-kubernetes\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758959 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-openvswitch\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.758984 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-netns\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.759011 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-script-lib\") pod \"7a31e12c-13a1-4052-b679-7115e1a3574e\" (UID: \"7a31e12c-13a1-4052-b679-7115e1a3574e\") " Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.759650 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.759687 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.759708 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760243 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-log-socket" (OuterVolumeSpecName: "log-socket") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760275 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760341 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760351 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760435 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760811 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760856 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-slash" (OuterVolumeSpecName: "host-slash") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760864 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-node-log" (OuterVolumeSpecName: "node-log") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760882 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760900 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.760916 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.761223 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.761261 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.761454 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.763748 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a31e12c-13a1-4052-b679-7115e1a3574e-kube-api-access-fhpnc" (OuterVolumeSpecName: "kube-api-access-fhpnc") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "kube-api-access-fhpnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.764658 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a31e12c-13a1-4052-b679-7115e1a3574e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.771920 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "7a31e12c-13a1-4052-b679-7115e1a3574e" (UID: "7a31e12c-13a1-4052-b679-7115e1a3574e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860411 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-ovn\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860460 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-slash\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860501 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-kubelet\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860525 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovnkube-script-lib\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860549 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-log-socket\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860567 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860589 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860626 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-node-log\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860644 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-cni-netd\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860669 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860744 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-systemd\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860766 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77jbq\" (UniqueName: \"kubernetes.io/projected/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-kube-api-access-77jbq\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860790 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-var-lib-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.860885 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-cni-bin\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861035 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovnkube-config\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861156 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovn-node-metrics-cert\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861247 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-env-overrides\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861325 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-systemd-units\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861476 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-etc-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861520 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-run-netns\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861569 4741 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861583 4741 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-systemd\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861595 4741 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-kubelet\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861606 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhpnc\" (UniqueName: \"kubernetes.io/projected/7a31e12c-13a1-4052-b679-7115e1a3574e-kube-api-access-fhpnc\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861618 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a31e12c-13a1-4052-b679-7115e1a3574e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861628 4741 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861636 4741 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861646 4741 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861654 4741 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861663 4741 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861671 4741 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-log-socket\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861679 4741 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861686 4741 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-node-log\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861694 4741 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-slash\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861701 4741 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861709 4741 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861717 4741 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-run-netns\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861724 4741 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a31e12c-13a1-4052-b679-7115e1a3574e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861732 4741 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-systemd-units\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.861739 4741 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a31e12c-13a1-4052-b679-7115e1a3574e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963107 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-systemd-units\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963202 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-etc-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963232 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-run-netns\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963248 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-systemd-units\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963271 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-ovn\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963329 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-etc-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963343 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-ovn\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963357 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-slash\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963381 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-run-netns\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963428 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-slash\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963546 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-kubelet\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963586 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovnkube-script-lib\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963623 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963646 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-log-socket\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963657 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-kubelet\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963679 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963697 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-log-socket\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963710 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963736 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-node-log\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963741 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963761 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-cni-netd\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963791 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963822 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963825 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-cni-netd\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963797 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-node-log\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963922 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-systemd\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963952 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77jbq\" (UniqueName: \"kubernetes.io/projected/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-kube-api-access-77jbq\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.963988 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-var-lib-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964016 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-run-systemd\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964015 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-cni-bin\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964047 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-host-cni-bin\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964127 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovnkube-config\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964123 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-var-lib-openvswitch\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964238 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovn-node-metrics-cert\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964296 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-env-overrides\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964839 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-env-overrides\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.964882 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovnkube-config\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.965363 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovnkube-script-lib\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.969879 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-ovn-node-metrics-cert\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:39 crc kubenswrapper[4741]: I0929 19:20:39.980154 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77jbq\" (UniqueName: \"kubernetes.io/projected/c98bfa2e-d1f2-4461-a57a-3f6493a9904f-kube-api-access-77jbq\") pod \"ovnkube-node-c4ps4\" (UID: \"c98bfa2e-d1f2-4461-a57a-3f6493a9904f\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.004480 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:40 crc kubenswrapper[4741]: W0929 19:20:40.021289 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc98bfa2e_d1f2_4461_a57a_3f6493a9904f.slice/crio-2bf8b286cf7cde23b8703cb909bf46609f816af9fa3da31833bc35f232f6d530 WatchSource:0}: Error finding container 2bf8b286cf7cde23b8703cb909bf46609f816af9fa3da31833bc35f232f6d530: Status 404 returned error can't find the container with id 2bf8b286cf7cde23b8703cb909bf46609f816af9fa3da31833bc35f232f6d530 Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.624851 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovn-acl-logging/0.log" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.625818 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mrdpd_7a31e12c-13a1-4052-b679-7115e1a3574e/ovn-controller/0.log" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.626510 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.626498 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mrdpd" event={"ID":"7a31e12c-13a1-4052-b679-7115e1a3574e","Type":"ContainerDied","Data":"befd95673e49e8c033b3abf7605e4fdd98f7b7816eee7b0cc17e022920e7390f"} Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.626734 4741 scope.go:117] "RemoveContainer" containerID="3637cdd09ff11428d033a392b90d98862680635742030c3645664937ffd7b6b2" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.629754 4741 generic.go:334] "Generic (PLEG): container finished" podID="c98bfa2e-d1f2-4461-a57a-3f6493a9904f" containerID="50177f79e819605e6b61ededb41671cb133db984f8f15556a2c7d21dc890e37a" exitCode=0 Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.629898 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerDied","Data":"50177f79e819605e6b61ededb41671cb133db984f8f15556a2c7d21dc890e37a"} Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.629950 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"2bf8b286cf7cde23b8703cb909bf46609f816af9fa3da31833bc35f232f6d530"} Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.633487 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/2.log" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.654248 4741 scope.go:117] "RemoveContainer" containerID="0001459fc9d22121b7626c26a38b5db75ece85ca47a7c6db599990a1cfc0c869" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.695280 4741 scope.go:117] "RemoveContainer" containerID="3d71d02cc401ff19dca0f79516990685c40d9bb21c5501cffddac8cd20ed38be" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.705102 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mrdpd"] Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.708776 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mrdpd"] Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.720147 4741 scope.go:117] "RemoveContainer" containerID="19b7759766a89bed6a4de1630c108b78aca9fe9e69384a3031bd549c32853e61" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.733185 4741 scope.go:117] "RemoveContainer" containerID="923d2d546ecb32b52d12cc259b3a2e14c44c5537dc57636d15eb68f373c3f814" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.753275 4741 scope.go:117] "RemoveContainer" containerID="68cd3993cf9c794c0e9d84bc726d46476a520092d46155b76460cdd6d50411e1" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.778589 4741 scope.go:117] "RemoveContainer" containerID="8d323c128a794bfff13deae865ed178ed0066e592ff746246108cb3d7a8a1520" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.813346 4741 scope.go:117] "RemoveContainer" containerID="d56920d32a21764e745d514deb038a11729b4f07b7e16f63b62f63ca32d731ff" Sep 29 19:20:40 crc kubenswrapper[4741]: I0929 19:20:40.849306 4741 scope.go:117] "RemoveContainer" containerID="1a030c5975f2bc9b669aecbf06788fe0494c00976e4dcf0e7aac1d449815bbc7" Sep 29 19:20:41 crc kubenswrapper[4741]: I0929 19:20:41.095453 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a31e12c-13a1-4052-b679-7115e1a3574e" path="/var/lib/kubelet/pods/7a31e12c-13a1-4052-b679-7115e1a3574e/volumes" Sep 29 19:20:41 crc kubenswrapper[4741]: I0929 19:20:41.645195 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"fa8c4a7b905e9ee6ae998e779524c44d6f38b49ab020cd545d4d185280b01e64"} Sep 29 19:20:41 crc kubenswrapper[4741]: I0929 19:20:41.645249 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"1999317a969ed38b02864938f81d7ca95dff7e8a092f9dceb9adc3196c212855"} Sep 29 19:20:41 crc kubenswrapper[4741]: I0929 19:20:41.645262 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"8409ec25a81732843859353aa58740e57cfa7427397334d3f0bad733aea47fb1"} Sep 29 19:20:41 crc kubenswrapper[4741]: I0929 19:20:41.645271 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"e15a9cd3696577524408ab97abcc7cb11747332d67dffa0fb0185efd9e6bfe33"} Sep 29 19:20:41 crc kubenswrapper[4741]: I0929 19:20:41.645279 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"7404d171aa20bc73c8049a48d1d9c86ad2c6a6701748c5f92d55b0b66318db74"} Sep 29 19:20:41 crc kubenswrapper[4741]: I0929 19:20:41.645287 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"fb01e8ff4f8466eae026ce139ce94b12658963e094285485958ba73f05621eb8"} Sep 29 19:20:43 crc kubenswrapper[4741]: I0929 19:20:43.663006 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"923a8dac57c277bb28ab78ea1fe3fb19d24f594ec887a2eb513d8e78faa78f39"} Sep 29 19:20:46 crc kubenswrapper[4741]: I0929 19:20:46.686297 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" event={"ID":"c98bfa2e-d1f2-4461-a57a-3f6493a9904f","Type":"ContainerStarted","Data":"8b3ec37549750362d688bfd1315139b317cd2864c166d2471a4db03b23368c7b"} Sep 29 19:20:46 crc kubenswrapper[4741]: I0929 19:20:46.688496 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:46 crc kubenswrapper[4741]: I0929 19:20:46.688604 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:46 crc kubenswrapper[4741]: I0929 19:20:46.715949 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" podStartSLOduration=7.71593293 podStartE2EDuration="7.71593293s" podCreationTimestamp="2025-09-29 19:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:20:46.712932838 +0000 UTC m=+688.360722180" watchObservedRunningTime="2025-09-29 19:20:46.71593293 +0000 UTC m=+688.363722262" Sep 29 19:20:46 crc kubenswrapper[4741]: I0929 19:20:46.722124 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.102871 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f"] Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.103801 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.106589 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.115807 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f"] Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.268569 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.268650 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.268761 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfhhs\" (UniqueName: \"kubernetes.io/projected/12a902af-c903-4004-9e1a-f3b4ab1cae5f-kube-api-access-qfhhs\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.370401 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.370495 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfhhs\" (UniqueName: \"kubernetes.io/projected/12a902af-c903-4004-9e1a-f3b4ab1cae5f-kube-api-access-qfhhs\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.370531 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.371051 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.371081 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.391494 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfhhs\" (UniqueName: \"kubernetes.io/projected/12a902af-c903-4004-9e1a-f3b4ab1cae5f-kube-api-access-qfhhs\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.455362 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.477671 4741 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(1797abef1e6599bbdb7a7f6611b3aa1568cedc2aeeccb31bbf03e04274e75893): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.477735 4741 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(1797abef1e6599bbdb7a7f6611b3aa1568cedc2aeeccb31bbf03e04274e75893): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.477757 4741 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(1797abef1e6599bbdb7a7f6611b3aa1568cedc2aeeccb31bbf03e04274e75893): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.477802 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace(12a902af-c903-4004-9e1a-f3b4ab1cae5f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace(12a902af-c903-4004-9e1a-f3b4ab1cae5f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(1797abef1e6599bbdb7a7f6611b3aa1568cedc2aeeccb31bbf03e04274e75893): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.692183 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.692475 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.693808 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: I0929 19:20:47.723466 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.727300 4741 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(d9580e7699bf35cad269e3dd418caa7deafb5327287b286f5865bb2434d49e1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.727416 4741 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(d9580e7699bf35cad269e3dd418caa7deafb5327287b286f5865bb2434d49e1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.727464 4741 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(d9580e7699bf35cad269e3dd418caa7deafb5327287b286f5865bb2434d49e1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:47 crc kubenswrapper[4741]: E0929 19:20:47.727546 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace(12a902af-c903-4004-9e1a-f3b4ab1cae5f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace(12a902af-c903-4004-9e1a-f3b4ab1cae5f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(d9580e7699bf35cad269e3dd418caa7deafb5327287b286f5865bb2434d49e1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" Sep 29 19:20:53 crc kubenswrapper[4741]: I0929 19:20:53.086943 4741 scope.go:117] "RemoveContainer" containerID="472e7bb64bf7085f0ba716a2fc6cc1eb722f462abeeec7bbfddd1fc9cfcc3941" Sep 29 19:20:53 crc kubenswrapper[4741]: E0929 19:20:53.088976 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-ql4kz_openshift-multus(565a367b-9a7a-4b93-b6f8-638b639fef9b)\"" pod="openshift-multus/multus-ql4kz" podUID="565a367b-9a7a-4b93-b6f8-638b639fef9b" Sep 29 19:20:59 crc kubenswrapper[4741]: I0929 19:20:59.091176 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:59 crc kubenswrapper[4741]: I0929 19:20:59.091844 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:59 crc kubenswrapper[4741]: E0929 19:20:59.123341 4741 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(b20e0185e0159c61ed3fbc8b6c7d8653eaa0c7014de13ab772b427eaa35b6e3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 19:20:59 crc kubenswrapper[4741]: E0929 19:20:59.123531 4741 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(b20e0185e0159c61ed3fbc8b6c7d8653eaa0c7014de13ab772b427eaa35b6e3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:59 crc kubenswrapper[4741]: E0929 19:20:59.123585 4741 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(b20e0185e0159c61ed3fbc8b6c7d8653eaa0c7014de13ab772b427eaa35b6e3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:20:59 crc kubenswrapper[4741]: E0929 19:20:59.123690 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace(12a902af-c903-4004-9e1a-f3b4ab1cae5f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace(12a902af-c903-4004-9e1a-f3b4ab1cae5f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_openshift-marketplace_12a902af-c903-4004-9e1a-f3b4ab1cae5f_0(b20e0185e0159c61ed3fbc8b6c7d8653eaa0c7014de13ab772b427eaa35b6e3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" Sep 29 19:21:07 crc kubenswrapper[4741]: I0929 19:21:07.086233 4741 scope.go:117] "RemoveContainer" containerID="472e7bb64bf7085f0ba716a2fc6cc1eb722f462abeeec7bbfddd1fc9cfcc3941" Sep 29 19:21:07 crc kubenswrapper[4741]: I0929 19:21:07.819065 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ql4kz_565a367b-9a7a-4b93-b6f8-638b639fef9b/kube-multus/2.log" Sep 29 19:21:07 crc kubenswrapper[4741]: I0929 19:21:07.819451 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ql4kz" event={"ID":"565a367b-9a7a-4b93-b6f8-638b639fef9b","Type":"ContainerStarted","Data":"bd28be94724837c234fa93c1ae5e171d59310f9fd94353d64e0065fa540c778b"} Sep 29 19:21:10 crc kubenswrapper[4741]: I0929 19:21:10.025236 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4ps4" Sep 29 19:21:14 crc kubenswrapper[4741]: I0929 19:21:14.085282 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:21:14 crc kubenswrapper[4741]: I0929 19:21:14.086279 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:21:14 crc kubenswrapper[4741]: I0929 19:21:14.306222 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f"] Sep 29 19:21:14 crc kubenswrapper[4741]: I0929 19:21:14.857028 4741 generic.go:334] "Generic (PLEG): container finished" podID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerID="82efd8682a86e1995b2e077f81f32ac5254880ab5abba561e5987c07439d6abb" exitCode=0 Sep 29 19:21:14 crc kubenswrapper[4741]: I0929 19:21:14.857067 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" event={"ID":"12a902af-c903-4004-9e1a-f3b4ab1cae5f","Type":"ContainerDied","Data":"82efd8682a86e1995b2e077f81f32ac5254880ab5abba561e5987c07439d6abb"} Sep 29 19:21:14 crc kubenswrapper[4741]: I0929 19:21:14.857098 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" event={"ID":"12a902af-c903-4004-9e1a-f3b4ab1cae5f","Type":"ContainerStarted","Data":"7882fdb81a1916e9ad4585c0d4c64c7fcf1d510be1aa54d091995e6a93fdad2f"} Sep 29 19:21:16 crc kubenswrapper[4741]: I0929 19:21:16.870375 4741 generic.go:334] "Generic (PLEG): container finished" podID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerID="7744c236d9b3c56c61b28de6c4cf6e9f342b70d43fb908429c72edf4e9c1b116" exitCode=0 Sep 29 19:21:16 crc kubenswrapper[4741]: I0929 19:21:16.870985 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" event={"ID":"12a902af-c903-4004-9e1a-f3b4ab1cae5f","Type":"ContainerDied","Data":"7744c236d9b3c56c61b28de6c4cf6e9f342b70d43fb908429c72edf4e9c1b116"} Sep 29 19:21:17 crc kubenswrapper[4741]: I0929 19:21:17.879241 4741 generic.go:334] "Generic (PLEG): container finished" podID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerID="0b3d11fac0e10bb97ebdd2ba7739501fd9108f2f72b99d6383998bff28a02e16" exitCode=0 Sep 29 19:21:17 crc kubenswrapper[4741]: I0929 19:21:17.879284 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" event={"ID":"12a902af-c903-4004-9e1a-f3b4ab1cae5f","Type":"ContainerDied","Data":"0b3d11fac0e10bb97ebdd2ba7739501fd9108f2f72b99d6383998bff28a02e16"} Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.096739 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.203464 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfhhs\" (UniqueName: \"kubernetes.io/projected/12a902af-c903-4004-9e1a-f3b4ab1cae5f-kube-api-access-qfhhs\") pod \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.203521 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-bundle\") pod \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.203571 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-util\") pod \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\" (UID: \"12a902af-c903-4004-9e1a-f3b4ab1cae5f\") " Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.204034 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-bundle" (OuterVolumeSpecName: "bundle") pod "12a902af-c903-4004-9e1a-f3b4ab1cae5f" (UID: "12a902af-c903-4004-9e1a-f3b4ab1cae5f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.209246 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12a902af-c903-4004-9e1a-f3b4ab1cae5f-kube-api-access-qfhhs" (OuterVolumeSpecName: "kube-api-access-qfhhs") pod "12a902af-c903-4004-9e1a-f3b4ab1cae5f" (UID: "12a902af-c903-4004-9e1a-f3b4ab1cae5f"). InnerVolumeSpecName "kube-api-access-qfhhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.216506 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-util" (OuterVolumeSpecName: "util") pod "12a902af-c903-4004-9e1a-f3b4ab1cae5f" (UID: "12a902af-c903-4004-9e1a-f3b4ab1cae5f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.304603 4741 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.304642 4741 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a902af-c903-4004-9e1a-f3b4ab1cae5f-util\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.304655 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfhhs\" (UniqueName: \"kubernetes.io/projected/12a902af-c903-4004-9e1a-f3b4ab1cae5f-kube-api-access-qfhhs\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.892035 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" event={"ID":"12a902af-c903-4004-9e1a-f3b4ab1cae5f","Type":"ContainerDied","Data":"7882fdb81a1916e9ad4585c0d4c64c7fcf1d510be1aa54d091995e6a93fdad2f"} Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.892339 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7882fdb81a1916e9ad4585c0d4c64c7fcf1d510be1aa54d091995e6a93fdad2f" Sep 29 19:21:19 crc kubenswrapper[4741]: I0929 19:21:19.892090 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.857746 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46"] Sep 29 19:21:23 crc kubenswrapper[4741]: E0929 19:21:23.858037 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerName="util" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.858057 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerName="util" Sep 29 19:21:23 crc kubenswrapper[4741]: E0929 19:21:23.858077 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerName="pull" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.858088 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerName="pull" Sep 29 19:21:23 crc kubenswrapper[4741]: E0929 19:21:23.858121 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerName="extract" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.858132 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerName="extract" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.858249 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a902af-c903-4004-9e1a-f3b4ab1cae5f" containerName="extract" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.858721 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.860960 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.861159 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-knx7x" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.861362 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.870032 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46"] Sep 29 19:21:23 crc kubenswrapper[4741]: I0929 19:21:23.966257 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmvll\" (UniqueName: \"kubernetes.io/projected/6567df3f-ff1b-4627-9b01-52eadd11d93c-kube-api-access-dmvll\") pod \"nmstate-operator-5d6f6cfd66-mfw46\" (UID: \"6567df3f-ff1b-4627-9b01-52eadd11d93c\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" Sep 29 19:21:24 crc kubenswrapper[4741]: I0929 19:21:24.067881 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmvll\" (UniqueName: \"kubernetes.io/projected/6567df3f-ff1b-4627-9b01-52eadd11d93c-kube-api-access-dmvll\") pod \"nmstate-operator-5d6f6cfd66-mfw46\" (UID: \"6567df3f-ff1b-4627-9b01-52eadd11d93c\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" Sep 29 19:21:24 crc kubenswrapper[4741]: I0929 19:21:24.085529 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmvll\" (UniqueName: \"kubernetes.io/projected/6567df3f-ff1b-4627-9b01-52eadd11d93c-kube-api-access-dmvll\") pod \"nmstate-operator-5d6f6cfd66-mfw46\" (UID: \"6567df3f-ff1b-4627-9b01-52eadd11d93c\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" Sep 29 19:21:24 crc kubenswrapper[4741]: I0929 19:21:24.181116 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" Sep 29 19:21:24 crc kubenswrapper[4741]: I0929 19:21:24.386703 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46"] Sep 29 19:21:24 crc kubenswrapper[4741]: W0929 19:21:24.394510 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6567df3f_ff1b_4627_9b01_52eadd11d93c.slice/crio-891b6b3df4f0f283d98906425e953c420d6018c19b6b23b2db9fa1b9e4815be6 WatchSource:0}: Error finding container 891b6b3df4f0f283d98906425e953c420d6018c19b6b23b2db9fa1b9e4815be6: Status 404 returned error can't find the container with id 891b6b3df4f0f283d98906425e953c420d6018c19b6b23b2db9fa1b9e4815be6 Sep 29 19:21:24 crc kubenswrapper[4741]: I0929 19:21:24.926688 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" event={"ID":"6567df3f-ff1b-4627-9b01-52eadd11d93c","Type":"ContainerStarted","Data":"891b6b3df4f0f283d98906425e953c420d6018c19b6b23b2db9fa1b9e4815be6"} Sep 29 19:21:26 crc kubenswrapper[4741]: I0929 19:21:26.941320 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" event={"ID":"6567df3f-ff1b-4627-9b01-52eadd11d93c","Type":"ContainerStarted","Data":"107bb24f44286a75f8e8d9f438a563e4fb679231aec2101f60e8a1c7ff139e0a"} Sep 29 19:21:26 crc kubenswrapper[4741]: I0929 19:21:26.965034 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-mfw46" podStartSLOduration=2.117535665 podStartE2EDuration="3.965014512s" podCreationTimestamp="2025-09-29 19:21:23 +0000 UTC" firstStartedPulling="2025-09-29 19:21:24.397342716 +0000 UTC m=+726.045132048" lastFinishedPulling="2025-09-29 19:21:26.244821563 +0000 UTC m=+727.892610895" observedRunningTime="2025-09-29 19:21:26.960879329 +0000 UTC m=+728.608668761" watchObservedRunningTime="2025-09-29 19:21:26.965014512 +0000 UTC m=+728.612803844" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.025459 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.026473 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.028235 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-26xh4" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.041404 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-c96rz"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.042641 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.046007 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.048223 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.072756 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-f8s2v"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.073610 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.078741 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-c96rz"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.122068 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh7mf\" (UniqueName: \"kubernetes.io/projected/6d63068c-acfa-46f8-a905-f21e7ff67e84-kube-api-access-kh7mf\") pod \"nmstate-webhook-6d689559c5-c96rz\" (UID: \"6d63068c-acfa-46f8-a905-f21e7ff67e84\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.122121 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2rtg\" (UniqueName: \"kubernetes.io/projected/e4b27a34-8d13-4f3e-97ec-f1428f02aef8-kube-api-access-t2rtg\") pod \"nmstate-metrics-58fcddf996-b6wcb\" (UID: \"e4b27a34-8d13-4f3e-97ec-f1428f02aef8\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.122148 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-nmstate-lock\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.122182 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-dbus-socket\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.122203 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbxh9\" (UniqueName: \"kubernetes.io/projected/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-kube-api-access-vbxh9\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.122368 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6d63068c-acfa-46f8-a905-f21e7ff67e84-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-c96rz\" (UID: \"6d63068c-acfa-46f8-a905-f21e7ff67e84\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.122439 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-ovs-socket\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.162722 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.163548 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.167525 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.169078 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.169293 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dg8nf" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.176242 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224732 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2rtg\" (UniqueName: \"kubernetes.io/projected/e4b27a34-8d13-4f3e-97ec-f1428f02aef8-kube-api-access-t2rtg\") pod \"nmstate-metrics-58fcddf996-b6wcb\" (UID: \"e4b27a34-8d13-4f3e-97ec-f1428f02aef8\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224785 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tj6q\" (UniqueName: \"kubernetes.io/projected/7111177c-0787-4b38-9858-82853f33807b-kube-api-access-2tj6q\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224808 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-nmstate-lock\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224828 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7111177c-0787-4b38-9858-82853f33807b-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224855 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-dbus-socket\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224875 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbxh9\" (UniqueName: \"kubernetes.io/projected/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-kube-api-access-vbxh9\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224890 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6d63068c-acfa-46f8-a905-f21e7ff67e84-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-c96rz\" (UID: \"6d63068c-acfa-46f8-a905-f21e7ff67e84\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224911 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-ovs-socket\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224937 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7111177c-0787-4b38-9858-82853f33807b-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.224975 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh7mf\" (UniqueName: \"kubernetes.io/projected/6d63068c-acfa-46f8-a905-f21e7ff67e84-kube-api-access-kh7mf\") pod \"nmstate-webhook-6d689559c5-c96rz\" (UID: \"6d63068c-acfa-46f8-a905-f21e7ff67e84\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.225419 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-nmstate-lock\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.225481 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-ovs-socket\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.225600 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-dbus-socket\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.230910 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6d63068c-acfa-46f8-a905-f21e7ff67e84-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-c96rz\" (UID: \"6d63068c-acfa-46f8-a905-f21e7ff67e84\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.241943 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh7mf\" (UniqueName: \"kubernetes.io/projected/6d63068c-acfa-46f8-a905-f21e7ff67e84-kube-api-access-kh7mf\") pod \"nmstate-webhook-6d689559c5-c96rz\" (UID: \"6d63068c-acfa-46f8-a905-f21e7ff67e84\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.244162 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbxh9\" (UniqueName: \"kubernetes.io/projected/6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5-kube-api-access-vbxh9\") pod \"nmstate-handler-f8s2v\" (UID: \"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5\") " pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.248995 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2rtg\" (UniqueName: \"kubernetes.io/projected/e4b27a34-8d13-4f3e-97ec-f1428f02aef8-kube-api-access-t2rtg\") pod \"nmstate-metrics-58fcddf996-b6wcb\" (UID: \"e4b27a34-8d13-4f3e-97ec-f1428f02aef8\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.326181 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7111177c-0787-4b38-9858-82853f33807b-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.326260 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tj6q\" (UniqueName: \"kubernetes.io/projected/7111177c-0787-4b38-9858-82853f33807b-kube-api-access-2tj6q\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.326284 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7111177c-0787-4b38-9858-82853f33807b-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.327116 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7111177c-0787-4b38-9858-82853f33807b-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: E0929 19:21:28.327198 4741 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Sep 29 19:21:28 crc kubenswrapper[4741]: E0929 19:21:28.327242 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7111177c-0787-4b38-9858-82853f33807b-plugin-serving-cert podName:7111177c-0787-4b38-9858-82853f33807b nodeName:}" failed. No retries permitted until 2025-09-29 19:21:28.827228642 +0000 UTC m=+730.475017964 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/7111177c-0787-4b38-9858-82853f33807b-plugin-serving-cert") pod "nmstate-console-plugin-864bb6dfb5-kcgmv" (UID: "7111177c-0787-4b38-9858-82853f33807b") : secret "plugin-serving-cert" not found Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.344878 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tj6q\" (UniqueName: \"kubernetes.io/projected/7111177c-0787-4b38-9858-82853f33807b-kube-api-access-2tj6q\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.357010 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.366011 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.389217 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.402293 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7ff8fb6fc8-czp79"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.402963 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.415128 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7ff8fb6fc8-czp79"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.427404 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-oauth-serving-cert\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.427441 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-service-ca\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.427470 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-oauth-config\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.427501 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-config\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.427518 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-trusted-ca-bundle\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.427547 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c54vw\" (UniqueName: \"kubernetes.io/projected/c809f6e7-292f-4259-a9b0-115a7b4f9001-kube-api-access-c54vw\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.427593 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-serving-cert\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.528191 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c54vw\" (UniqueName: \"kubernetes.io/projected/c809f6e7-292f-4259-a9b0-115a7b4f9001-kube-api-access-c54vw\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.528480 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-serving-cert\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.528511 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-oauth-serving-cert\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.528529 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-service-ca\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.528555 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-oauth-config\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.528587 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-config\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.528603 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-trusted-ca-bundle\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.529435 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-service-ca\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.530325 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-trusted-ca-bundle\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.530360 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-oauth-serving-cert\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.530575 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-config\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.532487 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-oauth-config\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.533251 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c809f6e7-292f-4259-a9b0-115a7b4f9001-console-serving-cert\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.542832 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c54vw\" (UniqueName: \"kubernetes.io/projected/c809f6e7-292f-4259-a9b0-115a7b4f9001-kube-api-access-c54vw\") pod \"console-7ff8fb6fc8-czp79\" (UID: \"c809f6e7-292f-4259-a9b0-115a7b4f9001\") " pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.720987 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.821298 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-c96rz"] Sep 29 19:21:28 crc kubenswrapper[4741]: W0929 19:21:28.824819 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d63068c_acfa_46f8_a905_f21e7ff67e84.slice/crio-87c170e7eb0671602f0858d47e75282962df92935b1db4a459c871e3e426197f WatchSource:0}: Error finding container 87c170e7eb0671602f0858d47e75282962df92935b1db4a459c871e3e426197f: Status 404 returned error can't find the container with id 87c170e7eb0671602f0858d47e75282962df92935b1db4a459c871e3e426197f Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.831052 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7111177c-0787-4b38-9858-82853f33807b-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.837964 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7111177c-0787-4b38-9858-82853f33807b-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-kcgmv\" (UID: \"7111177c-0787-4b38-9858-82853f33807b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.853647 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb"] Sep 29 19:21:28 crc kubenswrapper[4741]: W0929 19:21:28.905195 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4b27a34_8d13_4f3e_97ec_f1428f02aef8.slice/crio-ac4ea810bb7aff01aadae2b93aed0c4c527f023d830d5a67ec8914fc38fc5700 WatchSource:0}: Error finding container ac4ea810bb7aff01aadae2b93aed0c4c527f023d830d5a67ec8914fc38fc5700: Status 404 returned error can't find the container with id ac4ea810bb7aff01aadae2b93aed0c4c527f023d830d5a67ec8914fc38fc5700 Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.940901 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7ff8fb6fc8-czp79"] Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.953086 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" event={"ID":"e4b27a34-8d13-4f3e-97ec-f1428f02aef8","Type":"ContainerStarted","Data":"ac4ea810bb7aff01aadae2b93aed0c4c527f023d830d5a67ec8914fc38fc5700"} Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.953993 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" event={"ID":"6d63068c-acfa-46f8-a905-f21e7ff67e84","Type":"ContainerStarted","Data":"87c170e7eb0671602f0858d47e75282962df92935b1db4a459c871e3e426197f"} Sep 29 19:21:28 crc kubenswrapper[4741]: I0929 19:21:28.954966 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-f8s2v" event={"ID":"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5","Type":"ContainerStarted","Data":"96cf5f358774824b279402b9d77906e7dce2631cd407d1c5d0153f600a22cb0e"} Sep 29 19:21:29 crc kubenswrapper[4741]: I0929 19:21:29.076590 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" Sep 29 19:21:29 crc kubenswrapper[4741]: I0929 19:21:29.299228 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv"] Sep 29 19:21:29 crc kubenswrapper[4741]: I0929 19:21:29.964114 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7ff8fb6fc8-czp79" event={"ID":"c809f6e7-292f-4259-a9b0-115a7b4f9001","Type":"ContainerStarted","Data":"e059b7788fba0b400ada6dd45551fa96c0e188c95e216c602b171408e29c3ba0"} Sep 29 19:21:29 crc kubenswrapper[4741]: I0929 19:21:29.964159 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7ff8fb6fc8-czp79" event={"ID":"c809f6e7-292f-4259-a9b0-115a7b4f9001","Type":"ContainerStarted","Data":"fb007ba20111c217fa0829ae8e31d94d00c5d4351d9906af53947c18777896ba"} Sep 29 19:21:29 crc kubenswrapper[4741]: I0929 19:21:29.967153 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" event={"ID":"7111177c-0787-4b38-9858-82853f33807b","Type":"ContainerStarted","Data":"0e250f2a207874141350461c7399608e494db896b7ffafb2a917b552bb2523d7"} Sep 29 19:21:29 crc kubenswrapper[4741]: I0929 19:21:29.986615 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7ff8fb6fc8-czp79" podStartSLOduration=1.986593982 podStartE2EDuration="1.986593982s" podCreationTimestamp="2025-09-29 19:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:21:29.981655745 +0000 UTC m=+731.629445097" watchObservedRunningTime="2025-09-29 19:21:29.986593982 +0000 UTC m=+731.634383334" Sep 29 19:21:30 crc kubenswrapper[4741]: I0929 19:21:30.980170 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" event={"ID":"e4b27a34-8d13-4f3e-97ec-f1428f02aef8","Type":"ContainerStarted","Data":"50da0f78fe7f4fec9830e56000d61c13a3b05a466af04eebe29361cc6ffa9a56"} Sep 29 19:21:31 crc kubenswrapper[4741]: I0929 19:21:31.987071 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" event={"ID":"6d63068c-acfa-46f8-a905-f21e7ff67e84","Type":"ContainerStarted","Data":"f36d7bd63df02d0228a5a708c5c49669df0f794c9af7c250567f3c9524fd1725"} Sep 29 19:21:31 crc kubenswrapper[4741]: I0929 19:21:31.987614 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:31 crc kubenswrapper[4741]: I0929 19:21:31.988505 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-f8s2v" event={"ID":"6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5","Type":"ContainerStarted","Data":"1f7117193b894fe7b2a3ce881ba0f1b713d1e3e31d14e47c6717ec2d275126bb"} Sep 29 19:21:31 crc kubenswrapper[4741]: I0929 19:21:31.988599 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:31 crc kubenswrapper[4741]: I0929 19:21:31.989691 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" event={"ID":"7111177c-0787-4b38-9858-82853f33807b","Type":"ContainerStarted","Data":"c8e6182c5978eb3ebb033cc61f6809cdcd5504932b5a77ae0f6da79fe369c592"} Sep 29 19:21:32 crc kubenswrapper[4741]: I0929 19:21:32.007910 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" podStartSLOduration=1.882107006 podStartE2EDuration="4.007890939s" podCreationTimestamp="2025-09-29 19:21:28 +0000 UTC" firstStartedPulling="2025-09-29 19:21:28.832052864 +0000 UTC m=+730.479842186" lastFinishedPulling="2025-09-29 19:21:30.957836787 +0000 UTC m=+732.605626119" observedRunningTime="2025-09-29 19:21:32.003084056 +0000 UTC m=+733.650873388" watchObservedRunningTime="2025-09-29 19:21:32.007890939 +0000 UTC m=+733.655680271" Sep 29 19:21:32 crc kubenswrapper[4741]: I0929 19:21:32.015147 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-kcgmv" podStartSLOduration=1.567489208 podStartE2EDuration="4.015131851s" podCreationTimestamp="2025-09-29 19:21:28 +0000 UTC" firstStartedPulling="2025-09-29 19:21:29.309230901 +0000 UTC m=+730.957020233" lastFinishedPulling="2025-09-29 19:21:31.756873544 +0000 UTC m=+733.404662876" observedRunningTime="2025-09-29 19:21:32.014297464 +0000 UTC m=+733.662086806" watchObservedRunningTime="2025-09-29 19:21:32.015131851 +0000 UTC m=+733.662921183" Sep 29 19:21:32 crc kubenswrapper[4741]: I0929 19:21:32.033363 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-f8s2v" podStartSLOduration=1.698085539 podStartE2EDuration="4.033347773s" podCreationTimestamp="2025-09-29 19:21:28 +0000 UTC" firstStartedPulling="2025-09-29 19:21:28.447058769 +0000 UTC m=+730.094848101" lastFinishedPulling="2025-09-29 19:21:30.782320993 +0000 UTC m=+732.430110335" observedRunningTime="2025-09-29 19:21:32.030794481 +0000 UTC m=+733.678583833" watchObservedRunningTime="2025-09-29 19:21:32.033347773 +0000 UTC m=+733.681137105" Sep 29 19:21:34 crc kubenswrapper[4741]: I0929 19:21:34.004036 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" event={"ID":"e4b27a34-8d13-4f3e-97ec-f1428f02aef8","Type":"ContainerStarted","Data":"b9b3e7d812fea1d657f3bb8edf0efce61eb0714bb9dbeb9641800a0e688dbd65"} Sep 29 19:21:34 crc kubenswrapper[4741]: I0929 19:21:34.022269 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-b6wcb" podStartSLOduration=1.896842446 podStartE2EDuration="6.022255826s" podCreationTimestamp="2025-09-29 19:21:28 +0000 UTC" firstStartedPulling="2025-09-29 19:21:28.907843254 +0000 UTC m=+730.555632586" lastFinishedPulling="2025-09-29 19:21:33.033256634 +0000 UTC m=+734.681045966" observedRunningTime="2025-09-29 19:21:34.019708025 +0000 UTC m=+735.667497387" watchObservedRunningTime="2025-09-29 19:21:34.022255826 +0000 UTC m=+735.670045158" Sep 29 19:21:38 crc kubenswrapper[4741]: I0929 19:21:38.422315 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-f8s2v" Sep 29 19:21:38 crc kubenswrapper[4741]: I0929 19:21:38.721500 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:38 crc kubenswrapper[4741]: I0929 19:21:38.721557 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:38 crc kubenswrapper[4741]: I0929 19:21:38.726108 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:39 crc kubenswrapper[4741]: I0929 19:21:39.035279 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7ff8fb6fc8-czp79" Sep 29 19:21:39 crc kubenswrapper[4741]: I0929 19:21:39.081256 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-jd7m4"] Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.329937 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27cn9"] Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.430576 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6"] Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.430765 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" podUID="c71ee9d0-0827-4cac-aa2f-a6835259d37b" containerName="route-controller-manager" containerID="cri-o://78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345" gracePeriod=30 Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.771442 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.841170 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-client-ca\") pod \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.841915 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-client-ca" (OuterVolumeSpecName: "client-ca") pod "c71ee9d0-0827-4cac-aa2f-a6835259d37b" (UID: "c71ee9d0-0827-4cac-aa2f-a6835259d37b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.842127 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kws77\" (UniqueName: \"kubernetes.io/projected/c71ee9d0-0827-4cac-aa2f-a6835259d37b-kube-api-access-kws77\") pod \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.842818 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-config\") pod \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.842878 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71ee9d0-0827-4cac-aa2f-a6835259d37b-serving-cert\") pod \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\" (UID: \"c71ee9d0-0827-4cac-aa2f-a6835259d37b\") " Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.843154 4741 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.843548 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-config" (OuterVolumeSpecName: "config") pod "c71ee9d0-0827-4cac-aa2f-a6835259d37b" (UID: "c71ee9d0-0827-4cac-aa2f-a6835259d37b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.847073 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c71ee9d0-0827-4cac-aa2f-a6835259d37b-kube-api-access-kws77" (OuterVolumeSpecName: "kube-api-access-kws77") pod "c71ee9d0-0827-4cac-aa2f-a6835259d37b" (UID: "c71ee9d0-0827-4cac-aa2f-a6835259d37b"). InnerVolumeSpecName "kube-api-access-kws77". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.850595 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c71ee9d0-0827-4cac-aa2f-a6835259d37b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c71ee9d0-0827-4cac-aa2f-a6835259d37b" (UID: "c71ee9d0-0827-4cac-aa2f-a6835259d37b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.944091 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71ee9d0-0827-4cac-aa2f-a6835259d37b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.944129 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kws77\" (UniqueName: \"kubernetes.io/projected/c71ee9d0-0827-4cac-aa2f-a6835259d37b-kube-api-access-kws77\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:44 crc kubenswrapper[4741]: I0929 19:21:44.944142 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71ee9d0-0827-4cac-aa2f-a6835259d37b-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.065903 4741 generic.go:334] "Generic (PLEG): container finished" podID="c71ee9d0-0827-4cac-aa2f-a6835259d37b" containerID="78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345" exitCode=0 Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.065945 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" event={"ID":"c71ee9d0-0827-4cac-aa2f-a6835259d37b","Type":"ContainerDied","Data":"78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345"} Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.065963 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.065991 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6" event={"ID":"c71ee9d0-0827-4cac-aa2f-a6835259d37b","Type":"ContainerDied","Data":"9ec4ea6ac1589d01dbcdb1d1d1cc2f2da2e02bcd2ececa12b885c21778013644"} Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.066005 4741 scope.go:117] "RemoveContainer" containerID="78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.066049 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" podUID="ff8d9dff-2997-4c1c-9f8a-9f051247d11d" containerName="controller-manager" containerID="cri-o://06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2" gracePeriod=30 Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.084696 4741 scope.go:117] "RemoveContainer" containerID="78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345" Sep 29 19:21:45 crc kubenswrapper[4741]: E0929 19:21:45.085296 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345\": container with ID starting with 78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345 not found: ID does not exist" containerID="78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.085359 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345"} err="failed to get container status \"78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345\": rpc error: code = NotFound desc = could not find container \"78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345\": container with ID starting with 78e455e71eeda3b41ce9688e331e9c4bf6e67a9e694b42194603764304839345 not found: ID does not exist" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.105131 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6"] Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.107966 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4ttw6"] Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.343100 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.347698 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdmbc\" (UniqueName: \"kubernetes.io/projected/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-kube-api-access-vdmbc\") pod \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.347773 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-config\") pod \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.348089 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-serving-cert\") pod \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.348113 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-client-ca\") pod \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.348138 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-proxy-ca-bundles\") pod \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\" (UID: \"ff8d9dff-2997-4c1c-9f8a-9f051247d11d\") " Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.348703 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-client-ca" (OuterVolumeSpecName: "client-ca") pod "ff8d9dff-2997-4c1c-9f8a-9f051247d11d" (UID: "ff8d9dff-2997-4c1c-9f8a-9f051247d11d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.348727 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-config" (OuterVolumeSpecName: "config") pod "ff8d9dff-2997-4c1c-9f8a-9f051247d11d" (UID: "ff8d9dff-2997-4c1c-9f8a-9f051247d11d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.350736 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ff8d9dff-2997-4c1c-9f8a-9f051247d11d" (UID: "ff8d9dff-2997-4c1c-9f8a-9f051247d11d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.351274 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ff8d9dff-2997-4c1c-9f8a-9f051247d11d" (UID: "ff8d9dff-2997-4c1c-9f8a-9f051247d11d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.352807 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-kube-api-access-vdmbc" (OuterVolumeSpecName: "kube-api-access-vdmbc") pod "ff8d9dff-2997-4c1c-9f8a-9f051247d11d" (UID: "ff8d9dff-2997-4c1c-9f8a-9f051247d11d"). InnerVolumeSpecName "kube-api-access-vdmbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.449009 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.449043 4741 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.449052 4741 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.449062 4741 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.449072 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdmbc\" (UniqueName: \"kubernetes.io/projected/ff8d9dff-2997-4c1c-9f8a-9f051247d11d-kube-api-access-vdmbc\") on node \"crc\" DevicePath \"\"" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.561313 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9"] Sep 29 19:21:45 crc kubenswrapper[4741]: E0929 19:21:45.562001 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71ee9d0-0827-4cac-aa2f-a6835259d37b" containerName="route-controller-manager" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.562023 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71ee9d0-0827-4cac-aa2f-a6835259d37b" containerName="route-controller-manager" Sep 29 19:21:45 crc kubenswrapper[4741]: E0929 19:21:45.562043 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8d9dff-2997-4c1c-9f8a-9f051247d11d" containerName="controller-manager" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.562053 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8d9dff-2997-4c1c-9f8a-9f051247d11d" containerName="controller-manager" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.562193 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff8d9dff-2997-4c1c-9f8a-9f051247d11d" containerName="controller-manager" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.562221 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c71ee9d0-0827-4cac-aa2f-a6835259d37b" containerName="route-controller-manager" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.562713 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.564616 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b6954d848-c8bjj"] Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.565336 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.570326 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.570561 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.570691 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.570869 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.571027 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.571023 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.577084 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b6954d848-c8bjj"] Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.579572 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9"] Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652081 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5e4940-15a9-45ab-82f5-8aafd22fc263-config\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652123 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-client-ca\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652144 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v444g\" (UniqueName: \"kubernetes.io/projected/4f5e4940-15a9-45ab-82f5-8aafd22fc263-kube-api-access-v444g\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652172 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-serving-cert\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652318 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-config\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652343 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-proxy-ca-bundles\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652360 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f5e4940-15a9-45ab-82f5-8aafd22fc263-client-ca\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652436 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4fj9\" (UniqueName: \"kubernetes.io/projected/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-kube-api-access-h4fj9\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.652458 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f5e4940-15a9-45ab-82f5-8aafd22fc263-serving-cert\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753241 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v444g\" (UniqueName: \"kubernetes.io/projected/4f5e4940-15a9-45ab-82f5-8aafd22fc263-kube-api-access-v444g\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753299 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-serving-cert\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753332 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-config\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753353 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f5e4940-15a9-45ab-82f5-8aafd22fc263-client-ca\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753374 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-proxy-ca-bundles\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753422 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4fj9\" (UniqueName: \"kubernetes.io/projected/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-kube-api-access-h4fj9\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753451 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f5e4940-15a9-45ab-82f5-8aafd22fc263-serving-cert\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.753504 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5e4940-15a9-45ab-82f5-8aafd22fc263-config\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.754592 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-proxy-ca-bundles\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.754613 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-client-ca\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.755254 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f5e4940-15a9-45ab-82f5-8aafd22fc263-client-ca\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.755365 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5e4940-15a9-45ab-82f5-8aafd22fc263-config\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.755492 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-client-ca\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.755705 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-config\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.758043 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-serving-cert\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.765734 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f5e4940-15a9-45ab-82f5-8aafd22fc263-serving-cert\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.773333 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v444g\" (UniqueName: \"kubernetes.io/projected/4f5e4940-15a9-45ab-82f5-8aafd22fc263-kube-api-access-v444g\") pod \"route-controller-manager-5b477cfb48-jljt9\" (UID: \"4f5e4940-15a9-45ab-82f5-8aafd22fc263\") " pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.776905 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4fj9\" (UniqueName: \"kubernetes.io/projected/ccd093b3-b983-4bf9-b8fa-c27b53bd57c9-kube-api-access-h4fj9\") pod \"controller-manager-6b6954d848-c8bjj\" (UID: \"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9\") " pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.895642 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:45 crc kubenswrapper[4741]: I0929 19:21:45.909244 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.082421 4741 generic.go:334] "Generic (PLEG): container finished" podID="ff8d9dff-2997-4c1c-9f8a-9f051247d11d" containerID="06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2" exitCode=0 Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.082742 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" event={"ID":"ff8d9dff-2997-4c1c-9f8a-9f051247d11d","Type":"ContainerDied","Data":"06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2"} Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.082796 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" event={"ID":"ff8d9dff-2997-4c1c-9f8a-9f051247d11d","Type":"ContainerDied","Data":"15bc0bb0e855d419fbd1ceca12a79dde385ce53477a3a44f9b37fce057929ea8"} Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.082825 4741 scope.go:117] "RemoveContainer" containerID="06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2" Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.082963 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27cn9" Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.108152 4741 scope.go:117] "RemoveContainer" containerID="06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2" Sep 29 19:21:46 crc kubenswrapper[4741]: E0929 19:21:46.108824 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2\": container with ID starting with 06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2 not found: ID does not exist" containerID="06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2" Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.108856 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2"} err="failed to get container status \"06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2\": rpc error: code = NotFound desc = could not find container \"06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2\": container with ID starting with 06a3cc2153124f0a1dccbfd0e359d2ddc716ad3d39569f946f04f53967fac3e2 not found: ID does not exist" Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.125826 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27cn9"] Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.129413 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27cn9"] Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.162820 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b6954d848-c8bjj"] Sep 29 19:21:46 crc kubenswrapper[4741]: I0929 19:21:46.314407 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9"] Sep 29 19:21:46 crc kubenswrapper[4741]: W0929 19:21:46.329020 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f5e4940_15a9_45ab_82f5_8aafd22fc263.slice/crio-eb41615d7ca7dd4da62b17883892d9d3108b786fa9a1cfb64942dd4edf733777 WatchSource:0}: Error finding container eb41615d7ca7dd4da62b17883892d9d3108b786fa9a1cfb64942dd4edf733777: Status 404 returned error can't find the container with id eb41615d7ca7dd4da62b17883892d9d3108b786fa9a1cfb64942dd4edf733777 Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.092423 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c71ee9d0-0827-4cac-aa2f-a6835259d37b" path="/var/lib/kubelet/pods/c71ee9d0-0827-4cac-aa2f-a6835259d37b/volumes" Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.093344 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff8d9dff-2997-4c1c-9f8a-9f051247d11d" path="/var/lib/kubelet/pods/ff8d9dff-2997-4c1c-9f8a-9f051247d11d/volumes" Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.093802 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.093833 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.093844 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" event={"ID":"4f5e4940-15a9-45ab-82f5-8aafd22fc263","Type":"ContainerStarted","Data":"79c9941d7bff95ddaeda381a9d676d60b040608ce18474c6a02a156396ffa316"} Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.093858 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" event={"ID":"4f5e4940-15a9-45ab-82f5-8aafd22fc263","Type":"ContainerStarted","Data":"eb41615d7ca7dd4da62b17883892d9d3108b786fa9a1cfb64942dd4edf733777"} Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.093871 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" event={"ID":"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9","Type":"ContainerStarted","Data":"445280dd2fa061b0ffccab24c6e9f1d276d8e00db70f706e2be0774c3ba2080a"} Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.093880 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" event={"ID":"ccd093b3-b983-4bf9-b8fa-c27b53bd57c9","Type":"ContainerStarted","Data":"c6e6f108078397a611e0cf143fc27c1578c2948393a13c0ebada1252b5d99b0e"} Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.096273 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.097322 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.110051 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b477cfb48-jljt9" podStartSLOduration=3.110028636 podStartE2EDuration="3.110028636s" podCreationTimestamp="2025-09-29 19:21:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:21:47.10421037 +0000 UTC m=+748.751999702" watchObservedRunningTime="2025-09-29 19:21:47.110028636 +0000 UTC m=+748.757817988" Sep 29 19:21:47 crc kubenswrapper[4741]: I0929 19:21:47.155093 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b6954d848-c8bjj" podStartSLOduration=3.155060004 podStartE2EDuration="3.155060004s" podCreationTimestamp="2025-09-29 19:21:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:21:47.152250864 +0000 UTC m=+748.800040196" watchObservedRunningTime="2025-09-29 19:21:47.155060004 +0000 UTC m=+748.802849366" Sep 29 19:21:48 crc kubenswrapper[4741]: I0929 19:21:48.371788 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-c96rz" Sep 29 19:21:52 crc kubenswrapper[4741]: I0929 19:21:52.834107 4741 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.262986 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv"] Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.264456 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.272852 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.282952 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv"] Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.446097 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.446138 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8qw6\" (UniqueName: \"kubernetes.io/projected/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-kube-api-access-k8qw6\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.446176 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.546910 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.546946 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8qw6\" (UniqueName: \"kubernetes.io/projected/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-kube-api-access-k8qw6\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.546993 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.547359 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.547441 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.570142 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8qw6\" (UniqueName: \"kubernetes.io/projected/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-kube-api-access-k8qw6\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.583842 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:00 crc kubenswrapper[4741]: I0929 19:22:00.958242 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv"] Sep 29 19:22:01 crc kubenswrapper[4741]: I0929 19:22:01.189261 4741 generic.go:334] "Generic (PLEG): container finished" podID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerID="46d11445c4c79dc6240627d6c9bcc717cba58d12687ab2b268d80308fcd95384" exitCode=0 Sep 29 19:22:01 crc kubenswrapper[4741]: I0929 19:22:01.189323 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" event={"ID":"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d","Type":"ContainerDied","Data":"46d11445c4c79dc6240627d6c9bcc717cba58d12687ab2b268d80308fcd95384"} Sep 29 19:22:01 crc kubenswrapper[4741]: I0929 19:22:01.189591 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" event={"ID":"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d","Type":"ContainerStarted","Data":"1641fef6b365c787942422e3f85500771bfc722d6944646eaf25359d2b06b2de"} Sep 29 19:22:01 crc kubenswrapper[4741]: I0929 19:22:01.738916 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:22:01 crc kubenswrapper[4741]: I0929 19:22:01.738982 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.581954 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-46cwf"] Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.585763 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-46cwf"] Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.585921 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.776658 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-utilities\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.776751 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cztf2\" (UniqueName: \"kubernetes.io/projected/4b891663-7ee9-422f-89b6-9442f45dc10c-kube-api-access-cztf2\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.776870 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-catalog-content\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.878570 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cztf2\" (UniqueName: \"kubernetes.io/projected/4b891663-7ee9-422f-89b6-9442f45dc10c-kube-api-access-cztf2\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.878732 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-catalog-content\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.878794 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-utilities\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.879225 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-catalog-content\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.879250 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-utilities\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.897373 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cztf2\" (UniqueName: \"kubernetes.io/projected/4b891663-7ee9-422f-89b6-9442f45dc10c-kube-api-access-cztf2\") pod \"redhat-operators-46cwf\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:02 crc kubenswrapper[4741]: I0929 19:22:02.903679 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:03 crc kubenswrapper[4741]: I0929 19:22:03.202324 4741 generic.go:334] "Generic (PLEG): container finished" podID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerID="4113086fe41a1f569112dcf06f92e09b5a4e58b1993a5558a450f34859b5a0bd" exitCode=0 Sep 29 19:22:03 crc kubenswrapper[4741]: I0929 19:22:03.202518 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" event={"ID":"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d","Type":"ContainerDied","Data":"4113086fe41a1f569112dcf06f92e09b5a4e58b1993a5558a450f34859b5a0bd"} Sep 29 19:22:03 crc kubenswrapper[4741]: I0929 19:22:03.312487 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-46cwf"] Sep 29 19:22:03 crc kubenswrapper[4741]: W0929 19:22:03.318421 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b891663_7ee9_422f_89b6_9442f45dc10c.slice/crio-cad404907416d6f5b450e3f8af3f2f2993923430be856f23e8fd15d2ba52b2df WatchSource:0}: Error finding container cad404907416d6f5b450e3f8af3f2f2993923430be856f23e8fd15d2ba52b2df: Status 404 returned error can't find the container with id cad404907416d6f5b450e3f8af3f2f2993923430be856f23e8fd15d2ba52b2df Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.122451 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-jd7m4" podUID="932a4f81-a3d7-480f-a3bb-648b2e3bbd31" containerName="console" containerID="cri-o://fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb" gracePeriod=15 Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.212554 4741 generic.go:334] "Generic (PLEG): container finished" podID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerID="ee98ed8cc3b389d06bc9739bf301a9109d445cd7ab06246453b521089445c721" exitCode=0 Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.212620 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" event={"ID":"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d","Type":"ContainerDied","Data":"ee98ed8cc3b389d06bc9739bf301a9109d445cd7ab06246453b521089445c721"} Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.214909 4741 generic.go:334] "Generic (PLEG): container finished" podID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerID="836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb" exitCode=0 Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.214937 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46cwf" event={"ID":"4b891663-7ee9-422f-89b6-9442f45dc10c","Type":"ContainerDied","Data":"836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb"} Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.214953 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46cwf" event={"ID":"4b891663-7ee9-422f-89b6-9442f45dc10c","Type":"ContainerStarted","Data":"cad404907416d6f5b450e3f8af3f2f2993923430be856f23e8fd15d2ba52b2df"} Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.546327 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-jd7m4_932a4f81-a3d7-480f-a3bb-648b2e3bbd31/console/0.log" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.546399 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.618793 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-config\") pod \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.618848 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-serving-cert\") pod \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.618889 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-service-ca\") pod \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.618946 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr9sb\" (UniqueName: \"kubernetes.io/projected/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-kube-api-access-qr9sb\") pod \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.618978 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-oauth-serving-cert\") pod \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.619047 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-oauth-config\") pod \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.619133 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-trusted-ca-bundle\") pod \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\" (UID: \"932a4f81-a3d7-480f-a3bb-648b2e3bbd31\") " Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.619836 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-service-ca" (OuterVolumeSpecName: "service-ca") pod "932a4f81-a3d7-480f-a3bb-648b2e3bbd31" (UID: "932a4f81-a3d7-480f-a3bb-648b2e3bbd31"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.619879 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-config" (OuterVolumeSpecName: "console-config") pod "932a4f81-a3d7-480f-a3bb-648b2e3bbd31" (UID: "932a4f81-a3d7-480f-a3bb-648b2e3bbd31"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.619891 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "932a4f81-a3d7-480f-a3bb-648b2e3bbd31" (UID: "932a4f81-a3d7-480f-a3bb-648b2e3bbd31"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.619902 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "932a4f81-a3d7-480f-a3bb-648b2e3bbd31" (UID: "932a4f81-a3d7-480f-a3bb-648b2e3bbd31"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.620447 4741 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.620477 4741 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.620495 4741 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.620512 4741 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.624590 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-kube-api-access-qr9sb" (OuterVolumeSpecName: "kube-api-access-qr9sb") pod "932a4f81-a3d7-480f-a3bb-648b2e3bbd31" (UID: "932a4f81-a3d7-480f-a3bb-648b2e3bbd31"). InnerVolumeSpecName "kube-api-access-qr9sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.625040 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "932a4f81-a3d7-480f-a3bb-648b2e3bbd31" (UID: "932a4f81-a3d7-480f-a3bb-648b2e3bbd31"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.625690 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "932a4f81-a3d7-480f-a3bb-648b2e3bbd31" (UID: "932a4f81-a3d7-480f-a3bb-648b2e3bbd31"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.721131 4741 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.721167 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr9sb\" (UniqueName: \"kubernetes.io/projected/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-kube-api-access-qr9sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:04 crc kubenswrapper[4741]: I0929 19:22:04.721180 4741 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/932a4f81-a3d7-480f-a3bb-648b2e3bbd31-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.222347 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-jd7m4_932a4f81-a3d7-480f-a3bb-648b2e3bbd31/console/0.log" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.222670 4741 generic.go:334] "Generic (PLEG): container finished" podID="932a4f81-a3d7-480f-a3bb-648b2e3bbd31" containerID="fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb" exitCode=2 Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.222728 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jd7m4" event={"ID":"932a4f81-a3d7-480f-a3bb-648b2e3bbd31","Type":"ContainerDied","Data":"fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb"} Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.222790 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jd7m4" event={"ID":"932a4f81-a3d7-480f-a3bb-648b2e3bbd31","Type":"ContainerDied","Data":"c9f466d0c22fea5d859fcb3a534ef26a4d6abec61a5c377ab5aa7a93fca5c263"} Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.222818 4741 scope.go:117] "RemoveContainer" containerID="fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.222948 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jd7m4" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.224678 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46cwf" event={"ID":"4b891663-7ee9-422f-89b6-9442f45dc10c","Type":"ContainerStarted","Data":"cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b"} Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.243908 4741 scope.go:117] "RemoveContainer" containerID="fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb" Sep 29 19:22:05 crc kubenswrapper[4741]: E0929 19:22:05.244374 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb\": container with ID starting with fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb not found: ID does not exist" containerID="fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.244428 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb"} err="failed to get container status \"fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb\": rpc error: code = NotFound desc = could not find container \"fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb\": container with ID starting with fb567b26b3fb54c5bafce8e28a27ddfa4f6eb22bcc0aad1ab2d4a1c4be6af3fb not found: ID does not exist" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.244820 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-jd7m4"] Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.247794 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-jd7m4"] Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.554145 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.632696 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-bundle\") pod \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.632867 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-util\") pod \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.632979 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8qw6\" (UniqueName: \"kubernetes.io/projected/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-kube-api-access-k8qw6\") pod \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\" (UID: \"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d\") " Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.634873 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-bundle" (OuterVolumeSpecName: "bundle") pod "d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" (UID: "d33f5fb0-1f2e-43a2-8b7a-0061df4f821d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.636589 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-kube-api-access-k8qw6" (OuterVolumeSpecName: "kube-api-access-k8qw6") pod "d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" (UID: "d33f5fb0-1f2e-43a2-8b7a-0061df4f821d"). InnerVolumeSpecName "kube-api-access-k8qw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.646445 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-util" (OuterVolumeSpecName: "util") pod "d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" (UID: "d33f5fb0-1f2e-43a2-8b7a-0061df4f821d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.734737 4741 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-util\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.734782 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8qw6\" (UniqueName: \"kubernetes.io/projected/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-kube-api-access-k8qw6\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:05 crc kubenswrapper[4741]: I0929 19:22:05.734795 4741 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d33f5fb0-1f2e-43a2-8b7a-0061df4f821d-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:06 crc kubenswrapper[4741]: I0929 19:22:06.232712 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" event={"ID":"d33f5fb0-1f2e-43a2-8b7a-0061df4f821d","Type":"ContainerDied","Data":"1641fef6b365c787942422e3f85500771bfc722d6944646eaf25359d2b06b2de"} Sep 29 19:22:06 crc kubenswrapper[4741]: I0929 19:22:06.233677 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv" Sep 29 19:22:06 crc kubenswrapper[4741]: I0929 19:22:06.233724 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1641fef6b365c787942422e3f85500771bfc722d6944646eaf25359d2b06b2de" Sep 29 19:22:06 crc kubenswrapper[4741]: I0929 19:22:06.235520 4741 generic.go:334] "Generic (PLEG): container finished" podID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerID="cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b" exitCode=0 Sep 29 19:22:06 crc kubenswrapper[4741]: I0929 19:22:06.235556 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46cwf" event={"ID":"4b891663-7ee9-422f-89b6-9442f45dc10c","Type":"ContainerDied","Data":"cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b"} Sep 29 19:22:07 crc kubenswrapper[4741]: I0929 19:22:07.092920 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="932a4f81-a3d7-480f-a3bb-648b2e3bbd31" path="/var/lib/kubelet/pods/932a4f81-a3d7-480f-a3bb-648b2e3bbd31/volumes" Sep 29 19:22:07 crc kubenswrapper[4741]: I0929 19:22:07.244072 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46cwf" event={"ID":"4b891663-7ee9-422f-89b6-9442f45dc10c","Type":"ContainerStarted","Data":"b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d"} Sep 29 19:22:07 crc kubenswrapper[4741]: I0929 19:22:07.268960 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-46cwf" podStartSLOduration=2.799966322 podStartE2EDuration="5.268938996s" podCreationTimestamp="2025-09-29 19:22:02 +0000 UTC" firstStartedPulling="2025-09-29 19:22:04.216902623 +0000 UTC m=+765.864691945" lastFinishedPulling="2025-09-29 19:22:06.685875287 +0000 UTC m=+768.333664619" observedRunningTime="2025-09-29 19:22:07.267894494 +0000 UTC m=+768.915683826" watchObservedRunningTime="2025-09-29 19:22:07.268938996 +0000 UTC m=+768.916728348" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.769138 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7frrr"] Sep 29 19:22:12 crc kubenswrapper[4741]: E0929 19:22:12.769634 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerName="pull" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.769647 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerName="pull" Sep 29 19:22:12 crc kubenswrapper[4741]: E0929 19:22:12.769656 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932a4f81-a3d7-480f-a3bb-648b2e3bbd31" containerName="console" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.769662 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="932a4f81-a3d7-480f-a3bb-648b2e3bbd31" containerName="console" Sep 29 19:22:12 crc kubenswrapper[4741]: E0929 19:22:12.769674 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerName="extract" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.769680 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerName="extract" Sep 29 19:22:12 crc kubenswrapper[4741]: E0929 19:22:12.769690 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerName="util" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.769696 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerName="util" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.769785 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d33f5fb0-1f2e-43a2-8b7a-0061df4f821d" containerName="extract" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.769796 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="932a4f81-a3d7-480f-a3bb-648b2e3bbd31" containerName="console" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.770490 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.781634 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7frrr"] Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.904573 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.904635 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.921803 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-utilities\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.922145 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-catalog-content\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.922593 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv6jw\" (UniqueName: \"kubernetes.io/projected/05f4015a-db8a-49fc-8e48-18687c80f8bc-kube-api-access-fv6jw\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:12 crc kubenswrapper[4741]: I0929 19:22:12.946833 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.023540 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-utilities\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.023623 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-catalog-content\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.023645 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv6jw\" (UniqueName: \"kubernetes.io/projected/05f4015a-db8a-49fc-8e48-18687c80f8bc-kube-api-access-fv6jw\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.023993 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-utilities\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.024431 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-catalog-content\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.042891 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv6jw\" (UniqueName: \"kubernetes.io/projected/05f4015a-db8a-49fc-8e48-18687c80f8bc-kube-api-access-fv6jw\") pod \"redhat-marketplace-7frrr\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.088111 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.380253 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:13 crc kubenswrapper[4741]: I0929 19:22:13.484484 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7frrr"] Sep 29 19:22:14 crc kubenswrapper[4741]: I0929 19:22:14.285569 4741 generic.go:334] "Generic (PLEG): container finished" podID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerID="2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538" exitCode=0 Sep 29 19:22:14 crc kubenswrapper[4741]: I0929 19:22:14.285757 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7frrr" event={"ID":"05f4015a-db8a-49fc-8e48-18687c80f8bc","Type":"ContainerDied","Data":"2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538"} Sep 29 19:22:14 crc kubenswrapper[4741]: I0929 19:22:14.285823 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7frrr" event={"ID":"05f4015a-db8a-49fc-8e48-18687c80f8bc","Type":"ContainerStarted","Data":"3445bfe4f0d4da4fad622d96a01731248582e6d616beb43549111d070e1ce6d3"} Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.291303 4741 generic.go:334] "Generic (PLEG): container finished" podID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerID="1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0" exitCode=0 Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.291373 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7frrr" event={"ID":"05f4015a-db8a-49fc-8e48-18687c80f8bc","Type":"ContainerDied","Data":"1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0"} Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.367562 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-46cwf"] Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.367868 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-46cwf" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="registry-server" containerID="cri-o://b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d" gracePeriod=2 Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.847763 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.861945 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cztf2\" (UniqueName: \"kubernetes.io/projected/4b891663-7ee9-422f-89b6-9442f45dc10c-kube-api-access-cztf2\") pod \"4b891663-7ee9-422f-89b6-9442f45dc10c\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.862008 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-utilities\") pod \"4b891663-7ee9-422f-89b6-9442f45dc10c\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.862071 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-catalog-content\") pod \"4b891663-7ee9-422f-89b6-9442f45dc10c\" (UID: \"4b891663-7ee9-422f-89b6-9442f45dc10c\") " Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.862960 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-utilities" (OuterVolumeSpecName: "utilities") pod "4b891663-7ee9-422f-89b6-9442f45dc10c" (UID: "4b891663-7ee9-422f-89b6-9442f45dc10c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.883614 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b891663-7ee9-422f-89b6-9442f45dc10c-kube-api-access-cztf2" (OuterVolumeSpecName: "kube-api-access-cztf2") pod "4b891663-7ee9-422f-89b6-9442f45dc10c" (UID: "4b891663-7ee9-422f-89b6-9442f45dc10c"). InnerVolumeSpecName "kube-api-access-cztf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.963144 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cztf2\" (UniqueName: \"kubernetes.io/projected/4b891663-7ee9-422f-89b6-9442f45dc10c-kube-api-access-cztf2\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:15 crc kubenswrapper[4741]: I0929 19:22:15.963170 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.303056 4741 generic.go:334] "Generic (PLEG): container finished" podID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerID="b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d" exitCode=0 Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.303116 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46cwf" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.303155 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46cwf" event={"ID":"4b891663-7ee9-422f-89b6-9442f45dc10c","Type":"ContainerDied","Data":"b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d"} Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.303205 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46cwf" event={"ID":"4b891663-7ee9-422f-89b6-9442f45dc10c","Type":"ContainerDied","Data":"cad404907416d6f5b450e3f8af3f2f2993923430be856f23e8fd15d2ba52b2df"} Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.303229 4741 scope.go:117] "RemoveContainer" containerID="b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.305427 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7frrr" event={"ID":"05f4015a-db8a-49fc-8e48-18687c80f8bc","Type":"ContainerStarted","Data":"3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c"} Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.321774 4741 scope.go:117] "RemoveContainer" containerID="cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.339063 4741 scope.go:117] "RemoveContainer" containerID="836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.347614 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7frrr" podStartSLOduration=2.7918480629999998 podStartE2EDuration="4.347593953s" podCreationTimestamp="2025-09-29 19:22:12 +0000 UTC" firstStartedPulling="2025-09-29 19:22:14.28792352 +0000 UTC m=+775.935712852" lastFinishedPulling="2025-09-29 19:22:15.84366941 +0000 UTC m=+777.491458742" observedRunningTime="2025-09-29 19:22:16.344884296 +0000 UTC m=+777.992673628" watchObservedRunningTime="2025-09-29 19:22:16.347593953 +0000 UTC m=+777.995383275" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.364085 4741 scope.go:117] "RemoveContainer" containerID="b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d" Sep 29 19:22:16 crc kubenswrapper[4741]: E0929 19:22:16.364642 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d\": container with ID starting with b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d not found: ID does not exist" containerID="b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.364678 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d"} err="failed to get container status \"b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d\": rpc error: code = NotFound desc = could not find container \"b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d\": container with ID starting with b37d7f34fbf624e6130aa47bab3fe25e24740dcb98220f41924679e07b3f794d not found: ID does not exist" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.364704 4741 scope.go:117] "RemoveContainer" containerID="cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b" Sep 29 19:22:16 crc kubenswrapper[4741]: E0929 19:22:16.365244 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b\": container with ID starting with cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b not found: ID does not exist" containerID="cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.365274 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b"} err="failed to get container status \"cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b\": rpc error: code = NotFound desc = could not find container \"cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b\": container with ID starting with cefd17555eb4954f5e236207d725c1c147a053c5be2347c41953fcb6ae5e992b not found: ID does not exist" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.365293 4741 scope.go:117] "RemoveContainer" containerID="836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb" Sep 29 19:22:16 crc kubenswrapper[4741]: E0929 19:22:16.365598 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb\": container with ID starting with 836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb not found: ID does not exist" containerID="836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.365623 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb"} err="failed to get container status \"836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb\": rpc error: code = NotFound desc = could not find container \"836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb\": container with ID starting with 836bc271042550c0fb52dce321aa901e63e326117fbb7158e0b75e644e74acfb not found: ID does not exist" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.647212 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx"] Sep 29 19:22:16 crc kubenswrapper[4741]: E0929 19:22:16.647419 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="extract-content" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.647430 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="extract-content" Sep 29 19:22:16 crc kubenswrapper[4741]: E0929 19:22:16.647448 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="registry-server" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.647453 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="registry-server" Sep 29 19:22:16 crc kubenswrapper[4741]: E0929 19:22:16.647464 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="extract-utilities" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.647473 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="extract-utilities" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.647656 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" containerName="registry-server" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.648103 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.649727 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.649788 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.650622 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-6c57f" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.650896 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.651262 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.671822 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-webhook-cert\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.671877 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp9d2\" (UniqueName: \"kubernetes.io/projected/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-kube-api-access-qp9d2\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.671912 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-apiservice-cert\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.673751 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx"] Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.772636 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b891663-7ee9-422f-89b6-9442f45dc10c" (UID: "4b891663-7ee9-422f-89b6-9442f45dc10c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.772886 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-webhook-cert\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.772940 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp9d2\" (UniqueName: \"kubernetes.io/projected/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-kube-api-access-qp9d2\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.772975 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-apiservice-cert\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.773029 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b891663-7ee9-422f-89b6-9442f45dc10c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.780221 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-webhook-cert\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.782781 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-apiservice-cert\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.797097 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp9d2\" (UniqueName: \"kubernetes.io/projected/c8d5a915-940f-4fd2-9efd-7ee3a9dba397-kube-api-access-qp9d2\") pod \"metallb-operator-controller-manager-578897fd67-n2ssx\" (UID: \"c8d5a915-940f-4fd2-9efd-7ee3a9dba397\") " pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.898536 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6"] Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.899158 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.905982 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rjw7m" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.906048 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.913531 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.930448 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6"] Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.960263 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.961664 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-46cwf"] Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.965547 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-46cwf"] Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.976154 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czrzg\" (UniqueName: \"kubernetes.io/projected/9a35855e-0119-4a1f-928a-51a39f9aa326-kube-api-access-czrzg\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.976262 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a35855e-0119-4a1f-928a-51a39f9aa326-webhook-cert\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:16 crc kubenswrapper[4741]: I0929 19:22:16.976366 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a35855e-0119-4a1f-928a-51a39f9aa326-apiservice-cert\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.077709 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a35855e-0119-4a1f-928a-51a39f9aa326-apiservice-cert\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.077828 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czrzg\" (UniqueName: \"kubernetes.io/projected/9a35855e-0119-4a1f-928a-51a39f9aa326-kube-api-access-czrzg\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.077854 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a35855e-0119-4a1f-928a-51a39f9aa326-webhook-cert\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.082470 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a35855e-0119-4a1f-928a-51a39f9aa326-apiservice-cert\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.088124 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a35855e-0119-4a1f-928a-51a39f9aa326-webhook-cert\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.101241 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b891663-7ee9-422f-89b6-9442f45dc10c" path="/var/lib/kubelet/pods/4b891663-7ee9-422f-89b6-9442f45dc10c/volumes" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.118206 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czrzg\" (UniqueName: \"kubernetes.io/projected/9a35855e-0119-4a1f-928a-51a39f9aa326-kube-api-access-czrzg\") pod \"metallb-operator-webhook-server-68f987f9f5-zpvf6\" (UID: \"9a35855e-0119-4a1f-928a-51a39f9aa326\") " pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.214675 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.506711 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx"] Sep 29 19:22:17 crc kubenswrapper[4741]: W0929 19:22:17.515858 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8d5a915_940f_4fd2_9efd_7ee3a9dba397.slice/crio-1f1e50adceaa99847425b9f75b85428536f1a34502888c4d8edece9a24c402f2 WatchSource:0}: Error finding container 1f1e50adceaa99847425b9f75b85428536f1a34502888c4d8edece9a24c402f2: Status 404 returned error can't find the container with id 1f1e50adceaa99847425b9f75b85428536f1a34502888c4d8edece9a24c402f2 Sep 29 19:22:17 crc kubenswrapper[4741]: I0929 19:22:17.653557 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6"] Sep 29 19:22:17 crc kubenswrapper[4741]: W0929 19:22:17.659280 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a35855e_0119_4a1f_928a_51a39f9aa326.slice/crio-e1769c41eaf64bfdf8d14d13626c8e16b017c54277b66f6d5001ed51dcca746f WatchSource:0}: Error finding container e1769c41eaf64bfdf8d14d13626c8e16b017c54277b66f6d5001ed51dcca746f: Status 404 returned error can't find the container with id e1769c41eaf64bfdf8d14d13626c8e16b017c54277b66f6d5001ed51dcca746f Sep 29 19:22:18 crc kubenswrapper[4741]: I0929 19:22:18.330215 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" event={"ID":"c8d5a915-940f-4fd2-9efd-7ee3a9dba397","Type":"ContainerStarted","Data":"1f1e50adceaa99847425b9f75b85428536f1a34502888c4d8edece9a24c402f2"} Sep 29 19:22:18 crc kubenswrapper[4741]: I0929 19:22:18.331449 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" event={"ID":"9a35855e-0119-4a1f-928a-51a39f9aa326","Type":"ContainerStarted","Data":"e1769c41eaf64bfdf8d14d13626c8e16b017c54277b66f6d5001ed51dcca746f"} Sep 29 19:22:22 crc kubenswrapper[4741]: I0929 19:22:22.359920 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" event={"ID":"9a35855e-0119-4a1f-928a-51a39f9aa326","Type":"ContainerStarted","Data":"d656c1ef7dd36370f583c26b2dd021c5fab0343dbc559a45b0fd873ba579322d"} Sep 29 19:22:22 crc kubenswrapper[4741]: I0929 19:22:22.360336 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:22 crc kubenswrapper[4741]: I0929 19:22:22.361334 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" event={"ID":"c8d5a915-940f-4fd2-9efd-7ee3a9dba397","Type":"ContainerStarted","Data":"37a4839bae0b06f6c4dceaec60ea0a93c764123e2b49c0b0c6cee2a320f5b320"} Sep 29 19:22:22 crc kubenswrapper[4741]: I0929 19:22:22.361606 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:22 crc kubenswrapper[4741]: I0929 19:22:22.381120 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" podStartSLOduration=2.428425641 podStartE2EDuration="6.381104903s" podCreationTimestamp="2025-09-29 19:22:16 +0000 UTC" firstStartedPulling="2025-09-29 19:22:17.662809792 +0000 UTC m=+779.310599124" lastFinishedPulling="2025-09-29 19:22:21.615489054 +0000 UTC m=+783.263278386" observedRunningTime="2025-09-29 19:22:22.375916546 +0000 UTC m=+784.023705878" watchObservedRunningTime="2025-09-29 19:22:22.381104903 +0000 UTC m=+784.028894235" Sep 29 19:22:22 crc kubenswrapper[4741]: I0929 19:22:22.397474 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" podStartSLOduration=2.355371069 podStartE2EDuration="6.397459732s" podCreationTimestamp="2025-09-29 19:22:16 +0000 UTC" firstStartedPulling="2025-09-29 19:22:17.51898326 +0000 UTC m=+779.166772592" lastFinishedPulling="2025-09-29 19:22:21.561071923 +0000 UTC m=+783.208861255" observedRunningTime="2025-09-29 19:22:22.396581015 +0000 UTC m=+784.044370357" watchObservedRunningTime="2025-09-29 19:22:22.397459732 +0000 UTC m=+784.045249064" Sep 29 19:22:23 crc kubenswrapper[4741]: I0929 19:22:23.100349 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:23 crc kubenswrapper[4741]: I0929 19:22:23.100406 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:23 crc kubenswrapper[4741]: I0929 19:22:23.127370 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:23 crc kubenswrapper[4741]: I0929 19:22:23.407691 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:23 crc kubenswrapper[4741]: I0929 19:22:23.570821 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7frrr"] Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.377978 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7frrr" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="registry-server" containerID="cri-o://3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c" gracePeriod=2 Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.798260 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.896435 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-catalog-content\") pod \"05f4015a-db8a-49fc-8e48-18687c80f8bc\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.896542 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-utilities\") pod \"05f4015a-db8a-49fc-8e48-18687c80f8bc\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.896577 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv6jw\" (UniqueName: \"kubernetes.io/projected/05f4015a-db8a-49fc-8e48-18687c80f8bc-kube-api-access-fv6jw\") pod \"05f4015a-db8a-49fc-8e48-18687c80f8bc\" (UID: \"05f4015a-db8a-49fc-8e48-18687c80f8bc\") " Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.897345 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-utilities" (OuterVolumeSpecName: "utilities") pod "05f4015a-db8a-49fc-8e48-18687c80f8bc" (UID: "05f4015a-db8a-49fc-8e48-18687c80f8bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.909499 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05f4015a-db8a-49fc-8e48-18687c80f8bc" (UID: "05f4015a-db8a-49fc-8e48-18687c80f8bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.911107 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05f4015a-db8a-49fc-8e48-18687c80f8bc-kube-api-access-fv6jw" (OuterVolumeSpecName: "kube-api-access-fv6jw") pod "05f4015a-db8a-49fc-8e48-18687c80f8bc" (UID: "05f4015a-db8a-49fc-8e48-18687c80f8bc"). InnerVolumeSpecName "kube-api-access-fv6jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.998532 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.998561 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f4015a-db8a-49fc-8e48-18687c80f8bc-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:25 crc kubenswrapper[4741]: I0929 19:22:25.998570 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv6jw\" (UniqueName: \"kubernetes.io/projected/05f4015a-db8a-49fc-8e48-18687c80f8bc-kube-api-access-fv6jw\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.392408 4741 generic.go:334] "Generic (PLEG): container finished" podID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerID="3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c" exitCode=0 Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.392445 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7frrr" event={"ID":"05f4015a-db8a-49fc-8e48-18687c80f8bc","Type":"ContainerDied","Data":"3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c"} Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.392470 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7frrr" event={"ID":"05f4015a-db8a-49fc-8e48-18687c80f8bc","Type":"ContainerDied","Data":"3445bfe4f0d4da4fad622d96a01731248582e6d616beb43549111d070e1ce6d3"} Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.392486 4741 scope.go:117] "RemoveContainer" containerID="3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.392511 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7frrr" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.410653 4741 scope.go:117] "RemoveContainer" containerID="1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.424495 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7frrr"] Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.427636 4741 scope.go:117] "RemoveContainer" containerID="2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.431652 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7frrr"] Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.449081 4741 scope.go:117] "RemoveContainer" containerID="3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c" Sep 29 19:22:26 crc kubenswrapper[4741]: E0929 19:22:26.449874 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c\": container with ID starting with 3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c not found: ID does not exist" containerID="3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.449927 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c"} err="failed to get container status \"3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c\": rpc error: code = NotFound desc = could not find container \"3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c\": container with ID starting with 3d72f7732f7576fb12c18fe8897fc494b176913364af37be2147b4a3fd3ae07c not found: ID does not exist" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.449954 4741 scope.go:117] "RemoveContainer" containerID="1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0" Sep 29 19:22:26 crc kubenswrapper[4741]: E0929 19:22:26.450203 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0\": container with ID starting with 1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0 not found: ID does not exist" containerID="1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.450227 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0"} err="failed to get container status \"1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0\": rpc error: code = NotFound desc = could not find container \"1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0\": container with ID starting with 1849051435c42bdcbc164dabf1d5798be2faec68a7718dc0f4bd70fb63d327f0 not found: ID does not exist" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.450243 4741 scope.go:117] "RemoveContainer" containerID="2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538" Sep 29 19:22:26 crc kubenswrapper[4741]: E0929 19:22:26.450446 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538\": container with ID starting with 2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538 not found: ID does not exist" containerID="2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538" Sep 29 19:22:26 crc kubenswrapper[4741]: I0929 19:22:26.450467 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538"} err="failed to get container status \"2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538\": rpc error: code = NotFound desc = could not find container \"2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538\": container with ID starting with 2f6cdc70c14d759bfd6a8d131b7253162b754b889ae561b90567322d711ef538 not found: ID does not exist" Sep 29 19:22:27 crc kubenswrapper[4741]: I0929 19:22:27.092009 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" path="/var/lib/kubelet/pods/05f4015a-db8a-49fc-8e48-18687c80f8bc/volumes" Sep 29 19:22:31 crc kubenswrapper[4741]: I0929 19:22:31.739676 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:22:31 crc kubenswrapper[4741]: I0929 19:22:31.740333 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.773663 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ftg64"] Sep 29 19:22:32 crc kubenswrapper[4741]: E0929 19:22:32.773866 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="extract-utilities" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.773878 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="extract-utilities" Sep 29 19:22:32 crc kubenswrapper[4741]: E0929 19:22:32.773899 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="extract-content" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.773904 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="extract-content" Sep 29 19:22:32 crc kubenswrapper[4741]: E0929 19:22:32.773916 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="registry-server" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.773923 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="registry-server" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.774008 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="05f4015a-db8a-49fc-8e48-18687c80f8bc" containerName="registry-server" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.774724 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.802740 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ftg64"] Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.899030 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-catalog-content\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.899112 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-utilities\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:32 crc kubenswrapper[4741]: I0929 19:22:32.899195 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfg84\" (UniqueName: \"kubernetes.io/projected/5cb54e86-9399-4fea-bd58-4615a6b60be5-kube-api-access-nfg84\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.000590 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfg84\" (UniqueName: \"kubernetes.io/projected/5cb54e86-9399-4fea-bd58-4615a6b60be5-kube-api-access-nfg84\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.000637 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-catalog-content\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.000678 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-utilities\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.001161 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-utilities\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.001236 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-catalog-content\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.021670 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfg84\" (UniqueName: \"kubernetes.io/projected/5cb54e86-9399-4fea-bd58-4615a6b60be5-kube-api-access-nfg84\") pod \"community-operators-ftg64\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.095361 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.423988 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ftg64"] Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.775710 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f76qs"] Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.777871 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.787120 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f76qs"] Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.921718 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-catalog-content\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.922191 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw5lr\" (UniqueName: \"kubernetes.io/projected/e87eee9f-d51b-4112-9e78-c2bba7f5032d-kube-api-access-gw5lr\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:33 crc kubenswrapper[4741]: I0929 19:22:33.922273 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-utilities\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.023329 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-utilities\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.023492 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-catalog-content\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.023515 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw5lr\" (UniqueName: \"kubernetes.io/projected/e87eee9f-d51b-4112-9e78-c2bba7f5032d-kube-api-access-gw5lr\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.024050 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-catalog-content\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.024061 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-utilities\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.052570 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw5lr\" (UniqueName: \"kubernetes.io/projected/e87eee9f-d51b-4112-9e78-c2bba7f5032d-kube-api-access-gw5lr\") pod \"certified-operators-f76qs\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.105257 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.437533 4741 generic.go:334] "Generic (PLEG): container finished" podID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerID="e529d0b43a519ada2a2d7bd8fe613bbbeadec09e95298b0aa6c006fbd7c8112f" exitCode=0 Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.437647 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ftg64" event={"ID":"5cb54e86-9399-4fea-bd58-4615a6b60be5","Type":"ContainerDied","Data":"e529d0b43a519ada2a2d7bd8fe613bbbeadec09e95298b0aa6c006fbd7c8112f"} Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.438077 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ftg64" event={"ID":"5cb54e86-9399-4fea-bd58-4615a6b60be5","Type":"ContainerStarted","Data":"33ef49a1eb6a999cce56fd6a86d9a8a9478894c9e4d479ea776932d33c014426"} Sep 29 19:22:34 crc kubenswrapper[4741]: W0929 19:22:34.574901 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode87eee9f_d51b_4112_9e78_c2bba7f5032d.slice/crio-50e3ab85f8f5d2bf9343b69c01bc2060702e5dab12baf7478a3f32dcb26af296 WatchSource:0}: Error finding container 50e3ab85f8f5d2bf9343b69c01bc2060702e5dab12baf7478a3f32dcb26af296: Status 404 returned error can't find the container with id 50e3ab85f8f5d2bf9343b69c01bc2060702e5dab12baf7478a3f32dcb26af296 Sep 29 19:22:34 crc kubenswrapper[4741]: I0929 19:22:34.574913 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f76qs"] Sep 29 19:22:35 crc kubenswrapper[4741]: I0929 19:22:35.444258 4741 generic.go:334] "Generic (PLEG): container finished" podID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerID="80dc7df5c725f5795e4040a427e7c31d5f806caaa1574a67cc16ee9b34bf197a" exitCode=0 Sep 29 19:22:35 crc kubenswrapper[4741]: I0929 19:22:35.444361 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ftg64" event={"ID":"5cb54e86-9399-4fea-bd58-4615a6b60be5","Type":"ContainerDied","Data":"80dc7df5c725f5795e4040a427e7c31d5f806caaa1574a67cc16ee9b34bf197a"} Sep 29 19:22:35 crc kubenswrapper[4741]: I0929 19:22:35.446418 4741 generic.go:334] "Generic (PLEG): container finished" podID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerID="13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc" exitCode=0 Sep 29 19:22:35 crc kubenswrapper[4741]: I0929 19:22:35.446449 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f76qs" event={"ID":"e87eee9f-d51b-4112-9e78-c2bba7f5032d","Type":"ContainerDied","Data":"13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc"} Sep 29 19:22:35 crc kubenswrapper[4741]: I0929 19:22:35.446472 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f76qs" event={"ID":"e87eee9f-d51b-4112-9e78-c2bba7f5032d","Type":"ContainerStarted","Data":"50e3ab85f8f5d2bf9343b69c01bc2060702e5dab12baf7478a3f32dcb26af296"} Sep 29 19:22:36 crc kubenswrapper[4741]: I0929 19:22:36.455173 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ftg64" event={"ID":"5cb54e86-9399-4fea-bd58-4615a6b60be5","Type":"ContainerStarted","Data":"ef1f556a466f4de921b01f180615479105110ba665e49a125502cfea937cd3e0"} Sep 29 19:22:36 crc kubenswrapper[4741]: I0929 19:22:36.457010 4741 generic.go:334] "Generic (PLEG): container finished" podID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerID="8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9" exitCode=0 Sep 29 19:22:36 crc kubenswrapper[4741]: I0929 19:22:36.457073 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f76qs" event={"ID":"e87eee9f-d51b-4112-9e78-c2bba7f5032d","Type":"ContainerDied","Data":"8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9"} Sep 29 19:22:36 crc kubenswrapper[4741]: I0929 19:22:36.474147 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ftg64" podStartSLOduration=3.076318127 podStartE2EDuration="4.4741312s" podCreationTimestamp="2025-09-29 19:22:32 +0000 UTC" firstStartedPulling="2025-09-29 19:22:34.440043141 +0000 UTC m=+796.087832463" lastFinishedPulling="2025-09-29 19:22:35.837856204 +0000 UTC m=+797.485645536" observedRunningTime="2025-09-29 19:22:36.473674395 +0000 UTC m=+798.121463737" watchObservedRunningTime="2025-09-29 19:22:36.4741312 +0000 UTC m=+798.121920532" Sep 29 19:22:37 crc kubenswrapper[4741]: I0929 19:22:37.219213 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-68f987f9f5-zpvf6" Sep 29 19:22:37 crc kubenswrapper[4741]: I0929 19:22:37.464495 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f76qs" event={"ID":"e87eee9f-d51b-4112-9e78-c2bba7f5032d","Type":"ContainerStarted","Data":"d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685"} Sep 29 19:22:37 crc kubenswrapper[4741]: I0929 19:22:37.480955 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f76qs" podStartSLOduration=3.135280517 podStartE2EDuration="4.480938903s" podCreationTimestamp="2025-09-29 19:22:33 +0000 UTC" firstStartedPulling="2025-09-29 19:22:35.448276421 +0000 UTC m=+797.096065753" lastFinishedPulling="2025-09-29 19:22:36.793934807 +0000 UTC m=+798.441724139" observedRunningTime="2025-09-29 19:22:37.477831544 +0000 UTC m=+799.125620876" watchObservedRunningTime="2025-09-29 19:22:37.480938903 +0000 UTC m=+799.128728235" Sep 29 19:22:43 crc kubenswrapper[4741]: I0929 19:22:43.095587 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:43 crc kubenswrapper[4741]: I0929 19:22:43.096248 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:43 crc kubenswrapper[4741]: I0929 19:22:43.138893 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:43 crc kubenswrapper[4741]: I0929 19:22:43.539963 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:44 crc kubenswrapper[4741]: I0929 19:22:44.105999 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:44 crc kubenswrapper[4741]: I0929 19:22:44.106053 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:44 crc kubenswrapper[4741]: I0929 19:22:44.140057 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:44 crc kubenswrapper[4741]: I0929 19:22:44.561203 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:46 crc kubenswrapper[4741]: I0929 19:22:46.768227 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ftg64"] Sep 29 19:22:46 crc kubenswrapper[4741]: I0929 19:22:46.769038 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ftg64" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="registry-server" containerID="cri-o://ef1f556a466f4de921b01f180615479105110ba665e49a125502cfea937cd3e0" gracePeriod=2 Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.529616 4741 generic.go:334] "Generic (PLEG): container finished" podID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerID="ef1f556a466f4de921b01f180615479105110ba665e49a125502cfea937cd3e0" exitCode=0 Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.529684 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ftg64" event={"ID":"5cb54e86-9399-4fea-bd58-4615a6b60be5","Type":"ContainerDied","Data":"ef1f556a466f4de921b01f180615479105110ba665e49a125502cfea937cd3e0"} Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.697111 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.810321 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-utilities\") pod \"5cb54e86-9399-4fea-bd58-4615a6b60be5\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.810417 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfg84\" (UniqueName: \"kubernetes.io/projected/5cb54e86-9399-4fea-bd58-4615a6b60be5-kube-api-access-nfg84\") pod \"5cb54e86-9399-4fea-bd58-4615a6b60be5\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.810473 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-catalog-content\") pod \"5cb54e86-9399-4fea-bd58-4615a6b60be5\" (UID: \"5cb54e86-9399-4fea-bd58-4615a6b60be5\") " Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.811238 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-utilities" (OuterVolumeSpecName: "utilities") pod "5cb54e86-9399-4fea-bd58-4615a6b60be5" (UID: "5cb54e86-9399-4fea-bd58-4615a6b60be5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.815630 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb54e86-9399-4fea-bd58-4615a6b60be5-kube-api-access-nfg84" (OuterVolumeSpecName: "kube-api-access-nfg84") pod "5cb54e86-9399-4fea-bd58-4615a6b60be5" (UID: "5cb54e86-9399-4fea-bd58-4615a6b60be5"). InnerVolumeSpecName "kube-api-access-nfg84". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.853468 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cb54e86-9399-4fea-bd58-4615a6b60be5" (UID: "5cb54e86-9399-4fea-bd58-4615a6b60be5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.911716 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.911750 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfg84\" (UniqueName: \"kubernetes.io/projected/5cb54e86-9399-4fea-bd58-4615a6b60be5-kube-api-access-nfg84\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:47 crc kubenswrapper[4741]: I0929 19:22:47.911767 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb54e86-9399-4fea-bd58-4615a6b60be5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.537826 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ftg64" event={"ID":"5cb54e86-9399-4fea-bd58-4615a6b60be5","Type":"ContainerDied","Data":"33ef49a1eb6a999cce56fd6a86d9a8a9478894c9e4d479ea776932d33c014426"} Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.538150 4741 scope.go:117] "RemoveContainer" containerID="ef1f556a466f4de921b01f180615479105110ba665e49a125502cfea937cd3e0" Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.537901 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ftg64" Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.556679 4741 scope.go:117] "RemoveContainer" containerID="80dc7df5c725f5795e4040a427e7c31d5f806caaa1574a67cc16ee9b34bf197a" Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.568841 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ftg64"] Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.574917 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ftg64"] Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.595738 4741 scope.go:117] "RemoveContainer" containerID="e529d0b43a519ada2a2d7bd8fe613bbbeadec09e95298b0aa6c006fbd7c8112f" Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.768962 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f76qs"] Sep 29 19:22:48 crc kubenswrapper[4741]: I0929 19:22:48.769212 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f76qs" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="registry-server" containerID="cri-o://d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685" gracePeriod=2 Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.103162 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" path="/var/lib/kubelet/pods/5cb54e86-9399-4fea-bd58-4615a6b60be5/volumes" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.125378 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.234132 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-catalog-content\") pod \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.234210 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw5lr\" (UniqueName: \"kubernetes.io/projected/e87eee9f-d51b-4112-9e78-c2bba7f5032d-kube-api-access-gw5lr\") pod \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.234255 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-utilities\") pod \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\" (UID: \"e87eee9f-d51b-4112-9e78-c2bba7f5032d\") " Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.235300 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-utilities" (OuterVolumeSpecName: "utilities") pod "e87eee9f-d51b-4112-9e78-c2bba7f5032d" (UID: "e87eee9f-d51b-4112-9e78-c2bba7f5032d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.239138 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e87eee9f-d51b-4112-9e78-c2bba7f5032d-kube-api-access-gw5lr" (OuterVolumeSpecName: "kube-api-access-gw5lr") pod "e87eee9f-d51b-4112-9e78-c2bba7f5032d" (UID: "e87eee9f-d51b-4112-9e78-c2bba7f5032d"). InnerVolumeSpecName "kube-api-access-gw5lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.275708 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e87eee9f-d51b-4112-9e78-c2bba7f5032d" (UID: "e87eee9f-d51b-4112-9e78-c2bba7f5032d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.336079 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.336452 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw5lr\" (UniqueName: \"kubernetes.io/projected/e87eee9f-d51b-4112-9e78-c2bba7f5032d-kube-api-access-gw5lr\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.336536 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e87eee9f-d51b-4112-9e78-c2bba7f5032d-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.545813 4741 generic.go:334] "Generic (PLEG): container finished" podID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerID="d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685" exitCode=0 Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.545894 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f76qs" event={"ID":"e87eee9f-d51b-4112-9e78-c2bba7f5032d","Type":"ContainerDied","Data":"d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685"} Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.545926 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f76qs" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.546728 4741 scope.go:117] "RemoveContainer" containerID="d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.546918 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f76qs" event={"ID":"e87eee9f-d51b-4112-9e78-c2bba7f5032d","Type":"ContainerDied","Data":"50e3ab85f8f5d2bf9343b69c01bc2060702e5dab12baf7478a3f32dcb26af296"} Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.569024 4741 scope.go:117] "RemoveContainer" containerID="8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.617707 4741 scope.go:117] "RemoveContainer" containerID="13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.631320 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f76qs"] Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.634540 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f76qs"] Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.653659 4741 scope.go:117] "RemoveContainer" containerID="d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685" Sep 29 19:22:49 crc kubenswrapper[4741]: E0929 19:22:49.654028 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685\": container with ID starting with d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685 not found: ID does not exist" containerID="d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.654066 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685"} err="failed to get container status \"d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685\": rpc error: code = NotFound desc = could not find container \"d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685\": container with ID starting with d592e3fd0908d4786caf4cd94d89a25ade9ef3b1872781976c8d5fa21ee24685 not found: ID does not exist" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.654090 4741 scope.go:117] "RemoveContainer" containerID="8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9" Sep 29 19:22:49 crc kubenswrapper[4741]: E0929 19:22:49.654440 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9\": container with ID starting with 8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9 not found: ID does not exist" containerID="8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.654470 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9"} err="failed to get container status \"8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9\": rpc error: code = NotFound desc = could not find container \"8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9\": container with ID starting with 8b746a3bcd8a82dac258c9c06d4a787012aac05660c5c0c8a46fe4e6fba946d9 not found: ID does not exist" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.654490 4741 scope.go:117] "RemoveContainer" containerID="13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc" Sep 29 19:22:49 crc kubenswrapper[4741]: E0929 19:22:49.654783 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc\": container with ID starting with 13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc not found: ID does not exist" containerID="13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc" Sep 29 19:22:49 crc kubenswrapper[4741]: I0929 19:22:49.654810 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc"} err="failed to get container status \"13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc\": rpc error: code = NotFound desc = could not find container \"13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc\": container with ID starting with 13db620baba64c6d75bbf3d196434cbcc2cd6e81969b9a1b41a74f107f5687cc not found: ID does not exist" Sep 29 19:22:51 crc kubenswrapper[4741]: I0929 19:22:51.093295 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" path="/var/lib/kubelet/pods/e87eee9f-d51b-4112-9e78-c2bba7f5032d/volumes" Sep 29 19:22:56 crc kubenswrapper[4741]: I0929 19:22:56.963985 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-578897fd67-n2ssx" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.735154 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-2cfqh"] Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.735845 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="extract-utilities" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.735874 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="extract-utilities" Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.735905 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="registry-server" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.735919 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="registry-server" Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.735939 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="registry-server" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.735952 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="registry-server" Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.735971 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="extract-content" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.735983 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="extract-content" Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.735995 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="extract-content" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.736009 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="extract-content" Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.736034 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="extract-utilities" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.736046 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="extract-utilities" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.736224 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb54e86-9399-4fea-bd58-4615a6b60be5" containerName="registry-server" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.736255 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e87eee9f-d51b-4112-9e78-c2bba7f5032d" containerName="registry-server" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.739521 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.741903 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.744813 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.744862 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-m28s2" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.755766 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd"] Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.756441 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.760943 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.768167 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd"] Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840297 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-conf\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840340 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840360 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-startup\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840423 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics-certs\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840457 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-sockets\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840482 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6pb7\" (UniqueName: \"kubernetes.io/projected/bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5-kube-api-access-m6pb7\") pod \"frr-k8s-webhook-server-5478bdb765-hbchd\" (UID: \"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840554 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64wks\" (UniqueName: \"kubernetes.io/projected/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-kube-api-access-64wks\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840580 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-reloader\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.840604 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-hbchd\" (UID: \"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.844984 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-njmrg"] Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.845933 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-njmrg" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.847459 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.847608 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.847742 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.847799 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bfqw9" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.855484 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-lf5lv"] Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.856653 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.858363 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.874686 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-lf5lv"] Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941098 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6pb7\" (UniqueName: \"kubernetes.io/projected/bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5-kube-api-access-m6pb7\") pod \"frr-k8s-webhook-server-5478bdb765-hbchd\" (UID: \"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941148 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64wks\" (UniqueName: \"kubernetes.io/projected/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-kube-api-access-64wks\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941177 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-reloader\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941198 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bd8r\" (UniqueName: \"kubernetes.io/projected/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-kube-api-access-7bd8r\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941222 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-metrics-certs\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941238 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/80a22421-8515-4f7f-8af1-8cd2904154a0-metallb-excludel2\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941254 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-hbchd\" (UID: \"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941272 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941316 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-conf\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941330 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941452 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-startup\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941533 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-cert\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941556 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-metrics-certs\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941792 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics-certs\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.941935 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.941962 4741 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.942024 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-conf\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.942056 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-reloader\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.942605 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-startup\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: E0929 19:22:57.942842 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics-certs podName:c1a026f2-db7d-4ff5-b79a-ce82c325f71c nodeName:}" failed. No retries permitted until 2025-09-29 19:22:58.442825214 +0000 UTC m=+820.090614546 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics-certs") pod "frr-k8s-2cfqh" (UID: "c1a026f2-db7d-4ff5-b79a-ce82c325f71c") : secret "frr-k8s-certs-secret" not found Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.942885 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-sockets\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.943227 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-frr-sockets\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.942916 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n764l\" (UniqueName: \"kubernetes.io/projected/80a22421-8515-4f7f-8af1-8cd2904154a0-kube-api-access-n764l\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.950139 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-hbchd\" (UID: \"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.964462 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64wks\" (UniqueName: \"kubernetes.io/projected/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-kube-api-access-64wks\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:57 crc kubenswrapper[4741]: I0929 19:22:57.965129 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6pb7\" (UniqueName: \"kubernetes.io/projected/bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5-kube-api-access-m6pb7\") pod \"frr-k8s-webhook-server-5478bdb765-hbchd\" (UID: \"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.044762 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bd8r\" (UniqueName: \"kubernetes.io/projected/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-kube-api-access-7bd8r\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.045016 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-metrics-certs\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.045090 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/80a22421-8515-4f7f-8af1-8cd2904154a0-metallb-excludel2\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.045166 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.045262 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-cert\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.045326 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-metrics-certs\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.045444 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n764l\" (UniqueName: \"kubernetes.io/projected/80a22421-8515-4f7f-8af1-8cd2904154a0-kube-api-access-n764l\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: E0929 19:22:58.045250 4741 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 29 19:22:58 crc kubenswrapper[4741]: E0929 19:22:58.045793 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist podName:80a22421-8515-4f7f-8af1-8cd2904154a0 nodeName:}" failed. No retries permitted until 2025-09-29 19:22:58.545776975 +0000 UTC m=+820.193566297 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist") pod "speaker-njmrg" (UID: "80a22421-8515-4f7f-8af1-8cd2904154a0") : secret "metallb-memberlist" not found Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.046097 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/80a22421-8515-4f7f-8af1-8cd2904154a0-metallb-excludel2\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.048539 4741 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.048721 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-metrics-certs\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.050995 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-metrics-certs\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.058778 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-cert\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.073366 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bd8r\" (UniqueName: \"kubernetes.io/projected/b4d81ed4-fc0b-4f83-9207-2b6ab63ced26-kube-api-access-7bd8r\") pod \"controller-5d688f5ffc-lf5lv\" (UID: \"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26\") " pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.073994 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n764l\" (UniqueName: \"kubernetes.io/projected/80a22421-8515-4f7f-8af1-8cd2904154a0-kube-api-access-n764l\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.082828 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.178932 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.458662 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics-certs\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.464477 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1a026f2-db7d-4ff5-b79a-ce82c325f71c-metrics-certs\") pod \"frr-k8s-2cfqh\" (UID: \"c1a026f2-db7d-4ff5-b79a-ce82c325f71c\") " pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.560809 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:58 crc kubenswrapper[4741]: E0929 19:22:58.561002 4741 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 29 19:22:58 crc kubenswrapper[4741]: E0929 19:22:58.561095 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist podName:80a22421-8515-4f7f-8af1-8cd2904154a0 nodeName:}" failed. No retries permitted until 2025-09-29 19:22:59.561077396 +0000 UTC m=+821.208866728 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist") pod "speaker-njmrg" (UID: "80a22421-8515-4f7f-8af1-8cd2904154a0") : secret "metallb-memberlist" not found Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.594966 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd"] Sep 29 19:22:58 crc kubenswrapper[4741]: W0929 19:22:58.603345 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb1b55e6_2bf2_43c0_9c5f_c5896a59bac5.slice/crio-45ff428e8269b5b6f2504a0f56152cdaedb3fb2b7eeba66e0ca9d56aa9491845 WatchSource:0}: Error finding container 45ff428e8269b5b6f2504a0f56152cdaedb3fb2b7eeba66e0ca9d56aa9491845: Status 404 returned error can't find the container with id 45ff428e8269b5b6f2504a0f56152cdaedb3fb2b7eeba66e0ca9d56aa9491845 Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.646477 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-lf5lv"] Sep 29 19:22:58 crc kubenswrapper[4741]: I0929 19:22:58.669521 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.573227 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.592471 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/80a22421-8515-4f7f-8af1-8cd2904154a0-memberlist\") pod \"speaker-njmrg\" (UID: \"80a22421-8515-4f7f-8af1-8cd2904154a0\") " pod="metallb-system/speaker-njmrg" Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.611028 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerStarted","Data":"1df26a9399841f0024f9082d6e1af1438dc588ddc38d0ea1e9248f97439fcc13"} Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.612471 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" event={"ID":"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5","Type":"ContainerStarted","Data":"45ff428e8269b5b6f2504a0f56152cdaedb3fb2b7eeba66e0ca9d56aa9491845"} Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.614514 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-lf5lv" event={"ID":"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26","Type":"ContainerStarted","Data":"afc99a60d626b19f67cb434e57a4bff041f12d6466d5eeab0f4c5695b1cbd43f"} Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.614566 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-lf5lv" event={"ID":"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26","Type":"ContainerStarted","Data":"306a3122ebe193e69299bb85ecc173add4a9bff6c998745eaf0dd88280a8d7be"} Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.614581 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-lf5lv" event={"ID":"b4d81ed4-fc0b-4f83-9207-2b6ab63ced26","Type":"ContainerStarted","Data":"3986a2a6cddd3e002ac72090aa10a4f31e260d59232a347cd87f03efd6423520"} Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.614648 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.638830 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-lf5lv" podStartSLOduration=2.6388053339999997 podStartE2EDuration="2.638805334s" podCreationTimestamp="2025-09-29 19:22:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:22:59.635352252 +0000 UTC m=+821.283141614" watchObservedRunningTime="2025-09-29 19:22:59.638805334 +0000 UTC m=+821.286594666" Sep 29 19:22:59 crc kubenswrapper[4741]: I0929 19:22:59.660725 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-njmrg" Sep 29 19:23:00 crc kubenswrapper[4741]: I0929 19:23:00.639201 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-njmrg" event={"ID":"80a22421-8515-4f7f-8af1-8cd2904154a0","Type":"ContainerStarted","Data":"b5413b6a481ced93eae9aa0210c742d0588c397682390287189fbb91e644be2d"} Sep 29 19:23:00 crc kubenswrapper[4741]: I0929 19:23:00.639564 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-njmrg" event={"ID":"80a22421-8515-4f7f-8af1-8cd2904154a0","Type":"ContainerStarted","Data":"e8d6522bf4fafa854797a3e96030f7a8dc5a5b9bde00cf16f1c4c60befcd2350"} Sep 29 19:23:00 crc kubenswrapper[4741]: I0929 19:23:00.639586 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-njmrg" event={"ID":"80a22421-8515-4f7f-8af1-8cd2904154a0","Type":"ContainerStarted","Data":"ec18c3a720c35fdd8b5f62340682d0c6290131ea1f9e37900ee8f4c4e0eb2cc3"} Sep 29 19:23:00 crc kubenswrapper[4741]: I0929 19:23:00.639778 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-njmrg" Sep 29 19:23:00 crc kubenswrapper[4741]: I0929 19:23:00.660137 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-njmrg" podStartSLOduration=3.660119797 podStartE2EDuration="3.660119797s" podCreationTimestamp="2025-09-29 19:22:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:23:00.658889267 +0000 UTC m=+822.306678619" watchObservedRunningTime="2025-09-29 19:23:00.660119797 +0000 UTC m=+822.307909129" Sep 29 19:23:01 crc kubenswrapper[4741]: I0929 19:23:01.738832 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:23:01 crc kubenswrapper[4741]: I0929 19:23:01.738896 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:23:01 crc kubenswrapper[4741]: I0929 19:23:01.738944 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:23:01 crc kubenswrapper[4741]: I0929 19:23:01.739544 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f264b13b6ab5a53b78801880659a4525cf694d3e6348258f41be568cdac61c6"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:23:01 crc kubenswrapper[4741]: I0929 19:23:01.739595 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://1f264b13b6ab5a53b78801880659a4525cf694d3e6348258f41be568cdac61c6" gracePeriod=600 Sep 29 19:23:02 crc kubenswrapper[4741]: I0929 19:23:02.658221 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="1f264b13b6ab5a53b78801880659a4525cf694d3e6348258f41be568cdac61c6" exitCode=0 Sep 29 19:23:02 crc kubenswrapper[4741]: I0929 19:23:02.658284 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"1f264b13b6ab5a53b78801880659a4525cf694d3e6348258f41be568cdac61c6"} Sep 29 19:23:02 crc kubenswrapper[4741]: I0929 19:23:02.658565 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"d64910008ab65af756a5c6e2b068a8f502ccd9ab9a8a5bb1eac1cdbedf684c46"} Sep 29 19:23:02 crc kubenswrapper[4741]: I0929 19:23:02.658587 4741 scope.go:117] "RemoveContainer" containerID="fa3477edf9ba3614f07b8cc994bf0edeaf4d0348d0bf19dc3fa71df97308d735" Sep 29 19:23:05 crc kubenswrapper[4741]: I0929 19:23:05.678956 4741 generic.go:334] "Generic (PLEG): container finished" podID="c1a026f2-db7d-4ff5-b79a-ce82c325f71c" containerID="27ea361c7f140572815aeb193a539fd937e8750dbccb9ef643a6feea9c746204" exitCode=0 Sep 29 19:23:05 crc kubenswrapper[4741]: I0929 19:23:05.679078 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerDied","Data":"27ea361c7f140572815aeb193a539fd937e8750dbccb9ef643a6feea9c746204"} Sep 29 19:23:05 crc kubenswrapper[4741]: I0929 19:23:05.681184 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" event={"ID":"bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5","Type":"ContainerStarted","Data":"e17b3905731ea6bdf39acd2bccd39ace7b6e0c2e6196a1979346dfc249197f1e"} Sep 29 19:23:05 crc kubenswrapper[4741]: I0929 19:23:05.681342 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:23:06 crc kubenswrapper[4741]: I0929 19:23:06.687781 4741 generic.go:334] "Generic (PLEG): container finished" podID="c1a026f2-db7d-4ff5-b79a-ce82c325f71c" containerID="0c7843f4f91b534a4aa0e1f6f57de51e337a6bed2b5802b4c98c514bd70b9980" exitCode=0 Sep 29 19:23:06 crc kubenswrapper[4741]: I0929 19:23:06.687862 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerDied","Data":"0c7843f4f91b534a4aa0e1f6f57de51e337a6bed2b5802b4c98c514bd70b9980"} Sep 29 19:23:06 crc kubenswrapper[4741]: I0929 19:23:06.713481 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" podStartSLOduration=3.013984153 podStartE2EDuration="9.713462142s" podCreationTimestamp="2025-09-29 19:22:57 +0000 UTC" firstStartedPulling="2025-09-29 19:22:58.60664055 +0000 UTC m=+820.254429872" lastFinishedPulling="2025-09-29 19:23:05.306118519 +0000 UTC m=+826.953907861" observedRunningTime="2025-09-29 19:23:05.735893354 +0000 UTC m=+827.383682686" watchObservedRunningTime="2025-09-29 19:23:06.713462142 +0000 UTC m=+828.361251464" Sep 29 19:23:07 crc kubenswrapper[4741]: I0929 19:23:07.697783 4741 generic.go:334] "Generic (PLEG): container finished" podID="c1a026f2-db7d-4ff5-b79a-ce82c325f71c" containerID="abef227041c50ad9dc6929943af73c3e0e68d8abbea3c3176c1479c8df6d519c" exitCode=0 Sep 29 19:23:07 crc kubenswrapper[4741]: I0929 19:23:07.697825 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerDied","Data":"abef227041c50ad9dc6929943af73c3e0e68d8abbea3c3176c1479c8df6d519c"} Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.186383 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-lf5lv" Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.711279 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerStarted","Data":"899c17870f971112ca3e54b4a9cd6434789da5859b79ca114d51ba06e1d09e17"} Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.711546 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerStarted","Data":"7f42b76918eddb20e4751af4f94d3f5378ecceb7889e0f77d3eff71ebe5f554f"} Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.711591 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerStarted","Data":"5f52a8c28dea6b7cd129062cbe487e0555097d2931740d042abb3d4732e50a83"} Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.711620 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerStarted","Data":"c48f400e7d3ff74792e064306091445d147263519b82d262a4f50a9c353279f4"} Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.711633 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerStarted","Data":"05e31a32daf4d57c1979288d00dfb918cfe48cdbe55e0b3458a6263f3bf575a6"} Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.711643 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cfqh" event={"ID":"c1a026f2-db7d-4ff5-b79a-ce82c325f71c","Type":"ContainerStarted","Data":"d8f54172458397e46e6c01e97f1953b95de9981ada0953b04ee56eb52556840e"} Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.711698 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:23:08 crc kubenswrapper[4741]: I0929 19:23:08.731431 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-2cfqh" podStartSLOduration=5.225629607 podStartE2EDuration="11.73141121s" podCreationTimestamp="2025-09-29 19:22:57 +0000 UTC" firstStartedPulling="2025-09-29 19:22:58.783512422 +0000 UTC m=+820.431301754" lastFinishedPulling="2025-09-29 19:23:05.289294025 +0000 UTC m=+826.937083357" observedRunningTime="2025-09-29 19:23:08.730227772 +0000 UTC m=+830.378017134" watchObservedRunningTime="2025-09-29 19:23:08.73141121 +0000 UTC m=+830.379200542" Sep 29 19:23:09 crc kubenswrapper[4741]: I0929 19:23:09.667206 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-njmrg" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.194772 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd"] Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.196372 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.198530 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.207983 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd"] Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.345352 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77wjx\" (UniqueName: \"kubernetes.io/projected/7c0e9a8f-564e-472b-9f9f-fb571beeade4-kube-api-access-77wjx\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.345427 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.345503 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.447026 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77wjx\" (UniqueName: \"kubernetes.io/projected/7c0e9a8f-564e-472b-9f9f-fb571beeade4-kube-api-access-77wjx\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.447133 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.447222 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.447657 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.447671 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.469185 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77wjx\" (UniqueName: \"kubernetes.io/projected/7c0e9a8f-564e-472b-9f9f-fb571beeade4-kube-api-access-77wjx\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.510872 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:11 crc kubenswrapper[4741]: I0929 19:23:11.898931 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd"] Sep 29 19:23:11 crc kubenswrapper[4741]: W0929 19:23:11.903424 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c0e9a8f_564e_472b_9f9f_fb571beeade4.slice/crio-bd1536bb7799cfe7692bb34347c395a874d973a67fb202e73817d2092fa971b1 WatchSource:0}: Error finding container bd1536bb7799cfe7692bb34347c395a874d973a67fb202e73817d2092fa971b1: Status 404 returned error can't find the container with id bd1536bb7799cfe7692bb34347c395a874d973a67fb202e73817d2092fa971b1 Sep 29 19:23:12 crc kubenswrapper[4741]: I0929 19:23:12.746426 4741 generic.go:334] "Generic (PLEG): container finished" podID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerID="4705ca28d96ff31345cc83131719f425c4908901643ce44878030c35f8bef951" exitCode=0 Sep 29 19:23:12 crc kubenswrapper[4741]: I0929 19:23:12.746506 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" event={"ID":"7c0e9a8f-564e-472b-9f9f-fb571beeade4","Type":"ContainerDied","Data":"4705ca28d96ff31345cc83131719f425c4908901643ce44878030c35f8bef951"} Sep 29 19:23:12 crc kubenswrapper[4741]: I0929 19:23:12.746804 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" event={"ID":"7c0e9a8f-564e-472b-9f9f-fb571beeade4","Type":"ContainerStarted","Data":"bd1536bb7799cfe7692bb34347c395a874d973a67fb202e73817d2092fa971b1"} Sep 29 19:23:13 crc kubenswrapper[4741]: I0929 19:23:13.670747 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:23:13 crc kubenswrapper[4741]: I0929 19:23:13.730086 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:23:15 crc kubenswrapper[4741]: I0929 19:23:15.767334 4741 generic.go:334] "Generic (PLEG): container finished" podID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerID="07508d6b616ff35933e5821371e8dae4ba8ec1e963952babe06a87ef82079b7b" exitCode=0 Sep 29 19:23:15 crc kubenswrapper[4741]: I0929 19:23:15.767411 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" event={"ID":"7c0e9a8f-564e-472b-9f9f-fb571beeade4","Type":"ContainerDied","Data":"07508d6b616ff35933e5821371e8dae4ba8ec1e963952babe06a87ef82079b7b"} Sep 29 19:23:16 crc kubenswrapper[4741]: I0929 19:23:16.777517 4741 generic.go:334] "Generic (PLEG): container finished" podID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerID="24966d92b6ebdbdf80007026f0868537711d852a55ce21811074715e149cfb72" exitCode=0 Sep 29 19:23:16 crc kubenswrapper[4741]: I0929 19:23:16.777827 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" event={"ID":"7c0e9a8f-564e-472b-9f9f-fb571beeade4","Type":"ContainerDied","Data":"24966d92b6ebdbdf80007026f0868537711d852a55ce21811074715e149cfb72"} Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.057461 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.089517 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-hbchd" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.147625 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-util\") pod \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.147699 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77wjx\" (UniqueName: \"kubernetes.io/projected/7c0e9a8f-564e-472b-9f9f-fb571beeade4-kube-api-access-77wjx\") pod \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.147762 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-bundle\") pod \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\" (UID: \"7c0e9a8f-564e-472b-9f9f-fb571beeade4\") " Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.149060 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-bundle" (OuterVolumeSpecName: "bundle") pod "7c0e9a8f-564e-472b-9f9f-fb571beeade4" (UID: "7c0e9a8f-564e-472b-9f9f-fb571beeade4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.155040 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0e9a8f-564e-472b-9f9f-fb571beeade4-kube-api-access-77wjx" (OuterVolumeSpecName: "kube-api-access-77wjx") pod "7c0e9a8f-564e-472b-9f9f-fb571beeade4" (UID: "7c0e9a8f-564e-472b-9f9f-fb571beeade4"). InnerVolumeSpecName "kube-api-access-77wjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.169358 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-util" (OuterVolumeSpecName: "util") pod "7c0e9a8f-564e-472b-9f9f-fb571beeade4" (UID: "7c0e9a8f-564e-472b-9f9f-fb571beeade4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.249300 4741 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-util\") on node \"crc\" DevicePath \"\"" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.249348 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77wjx\" (UniqueName: \"kubernetes.io/projected/7c0e9a8f-564e-472b-9f9f-fb571beeade4-kube-api-access-77wjx\") on node \"crc\" DevicePath \"\"" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.249360 4741 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c0e9a8f-564e-472b-9f9f-fb571beeade4-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.673526 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-2cfqh" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.795007 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" event={"ID":"7c0e9a8f-564e-472b-9f9f-fb571beeade4","Type":"ContainerDied","Data":"bd1536bb7799cfe7692bb34347c395a874d973a67fb202e73817d2092fa971b1"} Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.795045 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd" Sep 29 19:23:18 crc kubenswrapper[4741]: I0929 19:23:18.795056 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd1536bb7799cfe7692bb34347c395a874d973a67fb202e73817d2092fa971b1" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.758770 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd"] Sep 29 19:23:24 crc kubenswrapper[4741]: E0929 19:23:24.759485 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerName="util" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.759497 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerName="util" Sep 29 19:23:24 crc kubenswrapper[4741]: E0929 19:23:24.759508 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerName="pull" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.759514 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerName="pull" Sep 29 19:23:24 crc kubenswrapper[4741]: E0929 19:23:24.759531 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerName="extract" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.759538 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerName="extract" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.759657 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0e9a8f-564e-472b-9f9f-fb571beeade4" containerName="extract" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.760073 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.762708 4741 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-4qkgv" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.766086 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.766632 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.773432 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd"] Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.845421 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cbsv\" (UniqueName: \"kubernetes.io/projected/6535437d-42a0-4a93-81d3-cf9cf82c278a-kube-api-access-4cbsv\") pod \"cert-manager-operator-controller-manager-57cd46d6d-wv5rd\" (UID: \"6535437d-42a0-4a93-81d3-cf9cf82c278a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.946495 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cbsv\" (UniqueName: \"kubernetes.io/projected/6535437d-42a0-4a93-81d3-cf9cf82c278a-kube-api-access-4cbsv\") pod \"cert-manager-operator-controller-manager-57cd46d6d-wv5rd\" (UID: \"6535437d-42a0-4a93-81d3-cf9cf82c278a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" Sep 29 19:23:24 crc kubenswrapper[4741]: I0929 19:23:24.972616 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cbsv\" (UniqueName: \"kubernetes.io/projected/6535437d-42a0-4a93-81d3-cf9cf82c278a-kube-api-access-4cbsv\") pod \"cert-manager-operator-controller-manager-57cd46d6d-wv5rd\" (UID: \"6535437d-42a0-4a93-81d3-cf9cf82c278a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" Sep 29 19:23:25 crc kubenswrapper[4741]: I0929 19:23:25.076128 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" Sep 29 19:23:25 crc kubenswrapper[4741]: I0929 19:23:25.460095 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd"] Sep 29 19:23:25 crc kubenswrapper[4741]: W0929 19:23:25.464222 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6535437d_42a0_4a93_81d3_cf9cf82c278a.slice/crio-275cf3bf1e001dbc0e49596ce12c561563ad77e8f663cfa6d097e5a0b3f6ea5c WatchSource:0}: Error finding container 275cf3bf1e001dbc0e49596ce12c561563ad77e8f663cfa6d097e5a0b3f6ea5c: Status 404 returned error can't find the container with id 275cf3bf1e001dbc0e49596ce12c561563ad77e8f663cfa6d097e5a0b3f6ea5c Sep 29 19:23:25 crc kubenswrapper[4741]: I0929 19:23:25.857492 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" event={"ID":"6535437d-42a0-4a93-81d3-cf9cf82c278a","Type":"ContainerStarted","Data":"275cf3bf1e001dbc0e49596ce12c561563ad77e8f663cfa6d097e5a0b3f6ea5c"} Sep 29 19:23:32 crc kubenswrapper[4741]: I0929 19:23:32.909630 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" event={"ID":"6535437d-42a0-4a93-81d3-cf9cf82c278a","Type":"ContainerStarted","Data":"4eea8c2932c54c524aa7fbc6755b03a274bccf3bbd43335fab837a8c886a0b11"} Sep 29 19:23:32 crc kubenswrapper[4741]: I0929 19:23:32.938589 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wv5rd" podStartSLOduration=2.230437023 podStartE2EDuration="8.938565628s" podCreationTimestamp="2025-09-29 19:23:24 +0000 UTC" firstStartedPulling="2025-09-29 19:23:25.471556159 +0000 UTC m=+847.119345491" lastFinishedPulling="2025-09-29 19:23:32.179684764 +0000 UTC m=+853.827474096" observedRunningTime="2025-09-29 19:23:32.936214723 +0000 UTC m=+854.584004095" watchObservedRunningTime="2025-09-29 19:23:32.938565628 +0000 UTC m=+854.586354990" Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.729717 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-gfzk9"] Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.730927 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.733130 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.733130 4741 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qwtzl" Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.733371 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.746617 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-gfzk9"] Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.900749 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddbc8c2b-e339-4850-aea0-094d3beaf857-bound-sa-token\") pod \"cert-manager-webhook-d969966f-gfzk9\" (UID: \"ddbc8c2b-e339-4850-aea0-094d3beaf857\") " pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:35 crc kubenswrapper[4741]: I0929 19:23:35.900848 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqqhw\" (UniqueName: \"kubernetes.io/projected/ddbc8c2b-e339-4850-aea0-094d3beaf857-kube-api-access-cqqhw\") pod \"cert-manager-webhook-d969966f-gfzk9\" (UID: \"ddbc8c2b-e339-4850-aea0-094d3beaf857\") " pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:36 crc kubenswrapper[4741]: I0929 19:23:36.003035 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqqhw\" (UniqueName: \"kubernetes.io/projected/ddbc8c2b-e339-4850-aea0-094d3beaf857-kube-api-access-cqqhw\") pod \"cert-manager-webhook-d969966f-gfzk9\" (UID: \"ddbc8c2b-e339-4850-aea0-094d3beaf857\") " pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:36 crc kubenswrapper[4741]: I0929 19:23:36.003198 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddbc8c2b-e339-4850-aea0-094d3beaf857-bound-sa-token\") pod \"cert-manager-webhook-d969966f-gfzk9\" (UID: \"ddbc8c2b-e339-4850-aea0-094d3beaf857\") " pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:36 crc kubenswrapper[4741]: I0929 19:23:36.025145 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqqhw\" (UniqueName: \"kubernetes.io/projected/ddbc8c2b-e339-4850-aea0-094d3beaf857-kube-api-access-cqqhw\") pod \"cert-manager-webhook-d969966f-gfzk9\" (UID: \"ddbc8c2b-e339-4850-aea0-094d3beaf857\") " pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:36 crc kubenswrapper[4741]: I0929 19:23:36.026255 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddbc8c2b-e339-4850-aea0-094d3beaf857-bound-sa-token\") pod \"cert-manager-webhook-d969966f-gfzk9\" (UID: \"ddbc8c2b-e339-4850-aea0-094d3beaf857\") " pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:36 crc kubenswrapper[4741]: I0929 19:23:36.048029 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:36 crc kubenswrapper[4741]: I0929 19:23:36.261437 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-gfzk9"] Sep 29 19:23:36 crc kubenswrapper[4741]: W0929 19:23:36.277636 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddbc8c2b_e339_4850_aea0_094d3beaf857.slice/crio-b5a98858f3079fb1fa08eba799f67fe1fdd6382853f0c08f6878474a2f9ec1ea WatchSource:0}: Error finding container b5a98858f3079fb1fa08eba799f67fe1fdd6382853f0c08f6878474a2f9ec1ea: Status 404 returned error can't find the container with id b5a98858f3079fb1fa08eba799f67fe1fdd6382853f0c08f6878474a2f9ec1ea Sep 29 19:23:36 crc kubenswrapper[4741]: I0929 19:23:36.936670 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" event={"ID":"ddbc8c2b-e339-4850-aea0-094d3beaf857","Type":"ContainerStarted","Data":"b5a98858f3079fb1fa08eba799f67fe1fdd6382853f0c08f6878474a2f9ec1ea"} Sep 29 19:23:39 crc kubenswrapper[4741]: I0929 19:23:39.899595 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf"] Sep 29 19:23:39 crc kubenswrapper[4741]: I0929 19:23:39.900896 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:39 crc kubenswrapper[4741]: I0929 19:23:39.903505 4741 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bbjx2" Sep 29 19:23:39 crc kubenswrapper[4741]: I0929 19:23:39.925624 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf"] Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.054354 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2dwv\" (UniqueName: \"kubernetes.io/projected/6eca0f9c-731e-49ea-ae29-40125ae4c0e2-kube-api-access-h2dwv\") pod \"cert-manager-cainjector-7d9f95dbf-g2rtf\" (UID: \"6eca0f9c-731e-49ea-ae29-40125ae4c0e2\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.054425 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6eca0f9c-731e-49ea-ae29-40125ae4c0e2-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-g2rtf\" (UID: \"6eca0f9c-731e-49ea-ae29-40125ae4c0e2\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.155746 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6eca0f9c-731e-49ea-ae29-40125ae4c0e2-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-g2rtf\" (UID: \"6eca0f9c-731e-49ea-ae29-40125ae4c0e2\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.155873 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2dwv\" (UniqueName: \"kubernetes.io/projected/6eca0f9c-731e-49ea-ae29-40125ae4c0e2-kube-api-access-h2dwv\") pod \"cert-manager-cainjector-7d9f95dbf-g2rtf\" (UID: \"6eca0f9c-731e-49ea-ae29-40125ae4c0e2\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.176021 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6eca0f9c-731e-49ea-ae29-40125ae4c0e2-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-g2rtf\" (UID: \"6eca0f9c-731e-49ea-ae29-40125ae4c0e2\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.176660 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2dwv\" (UniqueName: \"kubernetes.io/projected/6eca0f9c-731e-49ea-ae29-40125ae4c0e2-kube-api-access-h2dwv\") pod \"cert-manager-cainjector-7d9f95dbf-g2rtf\" (UID: \"6eca0f9c-731e-49ea-ae29-40125ae4c0e2\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.230650 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.851740 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf"] Sep 29 19:23:40 crc kubenswrapper[4741]: W0929 19:23:40.856219 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eca0f9c_731e_49ea_ae29_40125ae4c0e2.slice/crio-fe2064d7436cae39a31f160bfb487eb1ede0cc667e2afd8a7ebf7d270b03ba5a WatchSource:0}: Error finding container fe2064d7436cae39a31f160bfb487eb1ede0cc667e2afd8a7ebf7d270b03ba5a: Status 404 returned error can't find the container with id fe2064d7436cae39a31f160bfb487eb1ede0cc667e2afd8a7ebf7d270b03ba5a Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.964074 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" event={"ID":"ddbc8c2b-e339-4850-aea0-094d3beaf857","Type":"ContainerStarted","Data":"b0b6a1a7c07c04c2becff41a25f9882d82737a52af65da62e2792b45315c884a"} Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.964414 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.965800 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" event={"ID":"6eca0f9c-731e-49ea-ae29-40125ae4c0e2","Type":"ContainerStarted","Data":"27c7bd9d9d7c8cb8ae2e5dd281b3ab217e1f35376dd0765f7ce0d83f7d0129d5"} Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.965836 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" event={"ID":"6eca0f9c-731e-49ea-ae29-40125ae4c0e2","Type":"ContainerStarted","Data":"fe2064d7436cae39a31f160bfb487eb1ede0cc667e2afd8a7ebf7d270b03ba5a"} Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.997542 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-g2rtf" podStartSLOduration=1.99752423 podStartE2EDuration="1.99752423s" podCreationTimestamp="2025-09-29 19:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:23:40.994565667 +0000 UTC m=+862.642355009" watchObservedRunningTime="2025-09-29 19:23:40.99752423 +0000 UTC m=+862.645313562" Sep 29 19:23:40 crc kubenswrapper[4741]: I0929 19:23:40.998851 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" podStartSLOduration=1.7513588709999999 podStartE2EDuration="5.998841722s" podCreationTimestamp="2025-09-29 19:23:35 +0000 UTC" firstStartedPulling="2025-09-29 19:23:36.279732812 +0000 UTC m=+857.927522154" lastFinishedPulling="2025-09-29 19:23:40.527215673 +0000 UTC m=+862.175005005" observedRunningTime="2025-09-29 19:23:40.982495444 +0000 UTC m=+862.630284786" watchObservedRunningTime="2025-09-29 19:23:40.998841722 +0000 UTC m=+862.646631054" Sep 29 19:23:46 crc kubenswrapper[4741]: I0929 19:23:46.050659 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-gfzk9" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.527700 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-b9vcv"] Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.529040 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.533832 4741 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-s2p7t" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.548942 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-b9vcv"] Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.564976 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmjh4\" (UniqueName: \"kubernetes.io/projected/29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd-kube-api-access-zmjh4\") pod \"cert-manager-7d4cc89fcb-b9vcv\" (UID: \"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd\") " pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.565031 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-b9vcv\" (UID: \"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd\") " pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.666053 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmjh4\" (UniqueName: \"kubernetes.io/projected/29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd-kube-api-access-zmjh4\") pod \"cert-manager-7d4cc89fcb-b9vcv\" (UID: \"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd\") " pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.666112 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-b9vcv\" (UID: \"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd\") " pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.685530 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-b9vcv\" (UID: \"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd\") " pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.685845 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmjh4\" (UniqueName: \"kubernetes.io/projected/29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd-kube-api-access-zmjh4\") pod \"cert-manager-7d4cc89fcb-b9vcv\" (UID: \"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd\") " pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:54 crc kubenswrapper[4741]: I0929 19:23:54.857291 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" Sep 29 19:23:55 crc kubenswrapper[4741]: I0929 19:23:55.118830 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-b9vcv"] Sep 29 19:23:56 crc kubenswrapper[4741]: I0929 19:23:56.070960 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" event={"ID":"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd","Type":"ContainerStarted","Data":"071fb66b4786c63e3fe5b5b2ec9279c9f61e3823aaa7b890c5e82fd625aa6e1e"} Sep 29 19:23:59 crc kubenswrapper[4741]: I0929 19:23:59.096804 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" event={"ID":"29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd","Type":"ContainerStarted","Data":"b60dd4314fc0cbbe5d29987225e5b2b6a3d9bf112397c87c549bc45de5873fc9"} Sep 29 19:23:59 crc kubenswrapper[4741]: I0929 19:23:59.138546 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-b9vcv" podStartSLOduration=5.138519378 podStartE2EDuration="5.138519378s" podCreationTimestamp="2025-09-29 19:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:23:59.129198662 +0000 UTC m=+880.776988004" watchObservedRunningTime="2025-09-29 19:23:59.138519378 +0000 UTC m=+880.786308740" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.410222 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-w7bqm"] Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.411787 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.414069 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.414280 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.420802 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gjth2" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.428080 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w7bqm"] Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.573966 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d8h2\" (UniqueName: \"kubernetes.io/projected/3a0612c3-8758-4102-80e3-8082cef12fea-kube-api-access-6d8h2\") pod \"openstack-operator-index-w7bqm\" (UID: \"3a0612c3-8758-4102-80e3-8082cef12fea\") " pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.675974 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d8h2\" (UniqueName: \"kubernetes.io/projected/3a0612c3-8758-4102-80e3-8082cef12fea-kube-api-access-6d8h2\") pod \"openstack-operator-index-w7bqm\" (UID: \"3a0612c3-8758-4102-80e3-8082cef12fea\") " pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.706885 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d8h2\" (UniqueName: \"kubernetes.io/projected/3a0612c3-8758-4102-80e3-8082cef12fea-kube-api-access-6d8h2\") pod \"openstack-operator-index-w7bqm\" (UID: \"3a0612c3-8758-4102-80e3-8082cef12fea\") " pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:02 crc kubenswrapper[4741]: I0929 19:24:02.729975 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:03 crc kubenswrapper[4741]: I0929 19:24:03.176723 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w7bqm"] Sep 29 19:24:03 crc kubenswrapper[4741]: W0929 19:24:03.176773 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a0612c3_8758_4102_80e3_8082cef12fea.slice/crio-9f62e0e523fa60eae119f919e41856bb5f8c3ef1bf1c4a2ea4764abb8f3e068e WatchSource:0}: Error finding container 9f62e0e523fa60eae119f919e41856bb5f8c3ef1bf1c4a2ea4764abb8f3e068e: Status 404 returned error can't find the container with id 9f62e0e523fa60eae119f919e41856bb5f8c3ef1bf1c4a2ea4764abb8f3e068e Sep 29 19:24:04 crc kubenswrapper[4741]: I0929 19:24:04.129862 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w7bqm" event={"ID":"3a0612c3-8758-4102-80e3-8082cef12fea","Type":"ContainerStarted","Data":"9f62e0e523fa60eae119f919e41856bb5f8c3ef1bf1c4a2ea4764abb8f3e068e"} Sep 29 19:24:06 crc kubenswrapper[4741]: I0929 19:24:06.143470 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w7bqm" event={"ID":"3a0612c3-8758-4102-80e3-8082cef12fea","Type":"ContainerStarted","Data":"4eab7728cb61b1997b436a1ff4a34646ca2e9d3157b1c685d6d5c3439bd00929"} Sep 29 19:24:06 crc kubenswrapper[4741]: I0929 19:24:06.163671 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-w7bqm" podStartSLOduration=2.058457562 podStartE2EDuration="4.163647309s" podCreationTimestamp="2025-09-29 19:24:02 +0000 UTC" firstStartedPulling="2025-09-29 19:24:03.180337219 +0000 UTC m=+884.828126571" lastFinishedPulling="2025-09-29 19:24:05.285526986 +0000 UTC m=+886.933316318" observedRunningTime="2025-09-29 19:24:06.162590656 +0000 UTC m=+887.810380028" watchObservedRunningTime="2025-09-29 19:24:06.163647309 +0000 UTC m=+887.811436651" Sep 29 19:24:12 crc kubenswrapper[4741]: I0929 19:24:12.730493 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:12 crc kubenswrapper[4741]: I0929 19:24:12.730920 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:12 crc kubenswrapper[4741]: I0929 19:24:12.760980 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.212583 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-w7bqm" Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.824356 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27"] Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.825799 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.829647 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hkb8x" Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.834105 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27"] Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.925152 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-bundle\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.925213 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x57g\" (UniqueName: \"kubernetes.io/projected/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-kube-api-access-2x57g\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:13 crc kubenswrapper[4741]: I0929 19:24:13.925290 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-util\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.026116 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x57g\" (UniqueName: \"kubernetes.io/projected/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-kube-api-access-2x57g\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.026233 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-util\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.026266 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-bundle\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.026755 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-bundle\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.026926 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-util\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.045145 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x57g\" (UniqueName: \"kubernetes.io/projected/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-kube-api-access-2x57g\") pod \"36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.146988 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:14 crc kubenswrapper[4741]: I0929 19:24:14.329898 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27"] Sep 29 19:24:15 crc kubenswrapper[4741]: I0929 19:24:15.200156 4741 generic.go:334] "Generic (PLEG): container finished" podID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerID="f9f1e6ce341e037542d5fee6acf105ab6b74e3dcbda36ba5cc967661e445b82e" exitCode=0 Sep 29 19:24:15 crc kubenswrapper[4741]: I0929 19:24:15.200202 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" event={"ID":"b4f1fd1d-9a47-476a-8fda-ae251041b7dd","Type":"ContainerDied","Data":"f9f1e6ce341e037542d5fee6acf105ab6b74e3dcbda36ba5cc967661e445b82e"} Sep 29 19:24:15 crc kubenswrapper[4741]: I0929 19:24:15.200477 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" event={"ID":"b4f1fd1d-9a47-476a-8fda-ae251041b7dd","Type":"ContainerStarted","Data":"a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88"} Sep 29 19:24:16 crc kubenswrapper[4741]: I0929 19:24:16.206423 4741 generic.go:334] "Generic (PLEG): container finished" podID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerID="e63d8fee3523d5bef574907afe3f2d6add2cf9f9c3e8774ffd671e6438392425" exitCode=0 Sep 29 19:24:16 crc kubenswrapper[4741]: I0929 19:24:16.206459 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" event={"ID":"b4f1fd1d-9a47-476a-8fda-ae251041b7dd","Type":"ContainerDied","Data":"e63d8fee3523d5bef574907afe3f2d6add2cf9f9c3e8774ffd671e6438392425"} Sep 29 19:24:17 crc kubenswrapper[4741]: I0929 19:24:17.215187 4741 generic.go:334] "Generic (PLEG): container finished" podID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerID="21e075080ca88d2a69232a387eef62ae37d1a422e6604e101161d38a0559b18f" exitCode=0 Sep 29 19:24:17 crc kubenswrapper[4741]: I0929 19:24:17.215250 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" event={"ID":"b4f1fd1d-9a47-476a-8fda-ae251041b7dd","Type":"ContainerDied","Data":"21e075080ca88d2a69232a387eef62ae37d1a422e6604e101161d38a0559b18f"} Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.495783 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.584327 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-bundle\") pod \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.584381 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-util\") pod \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.584524 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x57g\" (UniqueName: \"kubernetes.io/projected/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-kube-api-access-2x57g\") pod \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\" (UID: \"b4f1fd1d-9a47-476a-8fda-ae251041b7dd\") " Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.585648 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-bundle" (OuterVolumeSpecName: "bundle") pod "b4f1fd1d-9a47-476a-8fda-ae251041b7dd" (UID: "b4f1fd1d-9a47-476a-8fda-ae251041b7dd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.591739 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-kube-api-access-2x57g" (OuterVolumeSpecName: "kube-api-access-2x57g") pod "b4f1fd1d-9a47-476a-8fda-ae251041b7dd" (UID: "b4f1fd1d-9a47-476a-8fda-ae251041b7dd"). InnerVolumeSpecName "kube-api-access-2x57g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.604240 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-util" (OuterVolumeSpecName: "util") pod "b4f1fd1d-9a47-476a-8fda-ae251041b7dd" (UID: "b4f1fd1d-9a47-476a-8fda-ae251041b7dd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.685824 4741 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.685884 4741 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-util\") on node \"crc\" DevicePath \"\"" Sep 29 19:24:18 crc kubenswrapper[4741]: I0929 19:24:18.685910 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x57g\" (UniqueName: \"kubernetes.io/projected/b4f1fd1d-9a47-476a-8fda-ae251041b7dd-kube-api-access-2x57g\") on node \"crc\" DevicePath \"\"" Sep 29 19:24:19 crc kubenswrapper[4741]: I0929 19:24:19.230925 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" event={"ID":"b4f1fd1d-9a47-476a-8fda-ae251041b7dd","Type":"ContainerDied","Data":"a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88"} Sep 29 19:24:19 crc kubenswrapper[4741]: I0929 19:24:19.231546 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88" Sep 29 19:24:19 crc kubenswrapper[4741]: I0929 19:24:19.230971 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27" Sep 29 19:24:19 crc kubenswrapper[4741]: E0929 19:24:19.530768 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice/crio-a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88\": RecentStats: unable to find data in memory cache]" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.876358 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj"] Sep 29 19:24:21 crc kubenswrapper[4741]: E0929 19:24:21.876887 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerName="extract" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.876899 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerName="extract" Sep 29 19:24:21 crc kubenswrapper[4741]: E0929 19:24:21.876922 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerName="pull" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.876928 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerName="pull" Sep 29 19:24:21 crc kubenswrapper[4741]: E0929 19:24:21.876936 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerName="util" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.876942 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerName="util" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.877034 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f1fd1d-9a47-476a-8fda-ae251041b7dd" containerName="extract" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.877627 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.880567 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5q2m2" Sep 29 19:24:21 crc kubenswrapper[4741]: I0929 19:24:21.901274 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj"] Sep 29 19:24:22 crc kubenswrapper[4741]: I0929 19:24:22.028074 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksj6x\" (UniqueName: \"kubernetes.io/projected/5ba00c60-cd9d-48ec-bbd2-e1202d32e161-kube-api-access-ksj6x\") pod \"openstack-operator-controller-operator-679cf76f6c-hvgtj\" (UID: \"5ba00c60-cd9d-48ec-bbd2-e1202d32e161\") " pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" Sep 29 19:24:22 crc kubenswrapper[4741]: I0929 19:24:22.129327 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksj6x\" (UniqueName: \"kubernetes.io/projected/5ba00c60-cd9d-48ec-bbd2-e1202d32e161-kube-api-access-ksj6x\") pod \"openstack-operator-controller-operator-679cf76f6c-hvgtj\" (UID: \"5ba00c60-cd9d-48ec-bbd2-e1202d32e161\") " pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" Sep 29 19:24:22 crc kubenswrapper[4741]: I0929 19:24:22.148949 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksj6x\" (UniqueName: \"kubernetes.io/projected/5ba00c60-cd9d-48ec-bbd2-e1202d32e161-kube-api-access-ksj6x\") pod \"openstack-operator-controller-operator-679cf76f6c-hvgtj\" (UID: \"5ba00c60-cd9d-48ec-bbd2-e1202d32e161\") " pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" Sep 29 19:24:22 crc kubenswrapper[4741]: I0929 19:24:22.193189 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" Sep 29 19:24:22 crc kubenswrapper[4741]: I0929 19:24:22.394732 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj"] Sep 29 19:24:22 crc kubenswrapper[4741]: W0929 19:24:22.401860 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba00c60_cd9d_48ec_bbd2_e1202d32e161.slice/crio-458303c9e00059aecba6b61fc2df1ccd16402abdee3b0fa6bac6df1c24e86904 WatchSource:0}: Error finding container 458303c9e00059aecba6b61fc2df1ccd16402abdee3b0fa6bac6df1c24e86904: Status 404 returned error can't find the container with id 458303c9e00059aecba6b61fc2df1ccd16402abdee3b0fa6bac6df1c24e86904 Sep 29 19:24:23 crc kubenswrapper[4741]: I0929 19:24:23.253749 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" event={"ID":"5ba00c60-cd9d-48ec-bbd2-e1202d32e161","Type":"ContainerStarted","Data":"458303c9e00059aecba6b61fc2df1ccd16402abdee3b0fa6bac6df1c24e86904"} Sep 29 19:24:26 crc kubenswrapper[4741]: I0929 19:24:26.275782 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" event={"ID":"5ba00c60-cd9d-48ec-bbd2-e1202d32e161","Type":"ContainerStarted","Data":"aab4f38761af02ecdaf0bf78a8601ffac126ee216a11d3b71190a26ba7c3ba8f"} Sep 29 19:24:28 crc kubenswrapper[4741]: I0929 19:24:28.288696 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" event={"ID":"5ba00c60-cd9d-48ec-bbd2-e1202d32e161","Type":"ContainerStarted","Data":"62bfc77d36028338fa62c6d99706d28a9e7f4244f10c75e0378bfd51be2fcd71"} Sep 29 19:24:28 crc kubenswrapper[4741]: I0929 19:24:28.289020 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" Sep 29 19:24:28 crc kubenswrapper[4741]: I0929 19:24:28.325569 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" podStartSLOduration=2.161694167 podStartE2EDuration="7.325551964s" podCreationTimestamp="2025-09-29 19:24:21 +0000 UTC" firstStartedPulling="2025-09-29 19:24:22.403734123 +0000 UTC m=+904.051523455" lastFinishedPulling="2025-09-29 19:24:27.56759193 +0000 UTC m=+909.215381252" observedRunningTime="2025-09-29 19:24:28.322048693 +0000 UTC m=+909.969838035" watchObservedRunningTime="2025-09-29 19:24:28.325551964 +0000 UTC m=+909.973341306" Sep 29 19:24:29 crc kubenswrapper[4741]: E0929 19:24:29.688083 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice/crio-a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice\": RecentStats: unable to find data in memory cache]" Sep 29 19:24:32 crc kubenswrapper[4741]: I0929 19:24:32.195945 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-679cf76f6c-hvgtj" Sep 29 19:24:39 crc kubenswrapper[4741]: E0929 19:24:39.835500 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice/crio-a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice\": RecentStats: unable to find data in memory cache]" Sep 29 19:24:49 crc kubenswrapper[4741]: E0929 19:24:49.984715 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice/crio-a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88\": RecentStats: unable to find data in memory cache]" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.336363 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.337808 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.347267 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-glw8s" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.349009 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.350740 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.364243 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.376230 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-z9lnx" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.391841 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.392701 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.398472 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.403899 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-gj9zw" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.412460 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.417182 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.418082 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.427221 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7v2r7" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.445624 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.450360 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.451313 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.458613 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-26wpt" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.475302 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrplq\" (UniqueName: \"kubernetes.io/projected/5637f538-f88c-4990-8e83-b40e3217c76d-kube-api-access-jrplq\") pod \"cinder-operator-controller-manager-644bddb6d8-dkh26\" (UID: \"5637f538-f88c-4990-8e83-b40e3217c76d\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.475368 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6hd4\" (UniqueName: \"kubernetes.io/projected/d3f19145-bf8f-4bb9-81cd-3fa20f066144-kube-api-access-l6hd4\") pod \"heat-operator-controller-manager-5d889d78cf-t5t9r\" (UID: \"d3f19145-bf8f-4bb9-81cd-3fa20f066144\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.475419 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr9r8\" (UniqueName: \"kubernetes.io/projected/2774171b-e907-4eee-a0cd-4f8c1871ad45-kube-api-access-sr9r8\") pod \"glance-operator-controller-manager-84958c4d49-vnpfp\" (UID: \"2774171b-e907-4eee-a0cd-4f8c1871ad45\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.475441 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbrdn\" (UniqueName: \"kubernetes.io/projected/00c10904-1ec2-4578-891e-2baa6ae5ceb7-kube-api-access-xbrdn\") pod \"designate-operator-controller-manager-84f4f7b77b-6d7tn\" (UID: \"00c10904-1ec2-4578-891e-2baa6ae5ceb7\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.475489 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s9gg\" (UniqueName: \"kubernetes.io/projected/9f3daf24-fa9f-48f9-9f0a-df998fa76551-kube-api-access-4s9gg\") pod \"barbican-operator-controller-manager-6ff8b75857-ntpxd\" (UID: \"9f3daf24-fa9f-48f9-9f0a-df998fa76551\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.481770 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.482806 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.487244 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-tj4vt" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.532927 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.536495 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.554535 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.555553 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.557368 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-b5qtm" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.577110 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6hd4\" (UniqueName: \"kubernetes.io/projected/d3f19145-bf8f-4bb9-81cd-3fa20f066144-kube-api-access-l6hd4\") pod \"heat-operator-controller-manager-5d889d78cf-t5t9r\" (UID: \"d3f19145-bf8f-4bb9-81cd-3fa20f066144\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.577156 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr9r8\" (UniqueName: \"kubernetes.io/projected/2774171b-e907-4eee-a0cd-4f8c1871ad45-kube-api-access-sr9r8\") pod \"glance-operator-controller-manager-84958c4d49-vnpfp\" (UID: \"2774171b-e907-4eee-a0cd-4f8c1871ad45\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.577181 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbrdn\" (UniqueName: \"kubernetes.io/projected/00c10904-1ec2-4578-891e-2baa6ae5ceb7-kube-api-access-xbrdn\") pod \"designate-operator-controller-manager-84f4f7b77b-6d7tn\" (UID: \"00c10904-1ec2-4578-891e-2baa6ae5ceb7\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.577220 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s9gg\" (UniqueName: \"kubernetes.io/projected/9f3daf24-fa9f-48f9-9f0a-df998fa76551-kube-api-access-4s9gg\") pod \"barbican-operator-controller-manager-6ff8b75857-ntpxd\" (UID: \"9f3daf24-fa9f-48f9-9f0a-df998fa76551\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.577263 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrplq\" (UniqueName: \"kubernetes.io/projected/5637f538-f88c-4990-8e83-b40e3217c76d-kube-api-access-jrplq\") pod \"cinder-operator-controller-manager-644bddb6d8-dkh26\" (UID: \"5637f538-f88c-4990-8e83-b40e3217c76d\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.585660 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.586642 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.603963 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-j6xnn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.604525 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.622504 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.667803 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6hd4\" (UniqueName: \"kubernetes.io/projected/d3f19145-bf8f-4bb9-81cd-3fa20f066144-kube-api-access-l6hd4\") pod \"heat-operator-controller-manager-5d889d78cf-t5t9r\" (UID: \"d3f19145-bf8f-4bb9-81cd-3fa20f066144\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.669546 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrplq\" (UniqueName: \"kubernetes.io/projected/5637f538-f88c-4990-8e83-b40e3217c76d-kube-api-access-jrplq\") pod \"cinder-operator-controller-manager-644bddb6d8-dkh26\" (UID: \"5637f538-f88c-4990-8e83-b40e3217c76d\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.670312 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s9gg\" (UniqueName: \"kubernetes.io/projected/9f3daf24-fa9f-48f9-9f0a-df998fa76551-kube-api-access-4s9gg\") pod \"barbican-operator-controller-manager-6ff8b75857-ntpxd\" (UID: \"9f3daf24-fa9f-48f9-9f0a-df998fa76551\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.677811 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbrdn\" (UniqueName: \"kubernetes.io/projected/00c10904-1ec2-4578-891e-2baa6ae5ceb7-kube-api-access-xbrdn\") pod \"designate-operator-controller-manager-84f4f7b77b-6d7tn\" (UID: \"00c10904-1ec2-4578-891e-2baa6ae5ceb7\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.717183 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6pdg\" (UniqueName: \"kubernetes.io/projected/a581380e-2a13-42a4-8a12-606c842cefbc-kube-api-access-v6pdg\") pod \"horizon-operator-controller-manager-9f4696d94-zd6mn\" (UID: \"a581380e-2a13-42a4-8a12-606c842cefbc\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.717282 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ckwb\" (UniqueName: \"kubernetes.io/projected/836c126e-917e-4b62-a2aa-7d4caac672e4-kube-api-access-7ckwb\") pod \"ironic-operator-controller-manager-7975b88857-f44nr\" (UID: \"836c126e-917e-4b62-a2aa-7d4caac672e4\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.719171 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr9r8\" (UniqueName: \"kubernetes.io/projected/2774171b-e907-4eee-a0cd-4f8c1871ad45-kube-api-access-sr9r8\") pod \"glance-operator-controller-manager-84958c4d49-vnpfp\" (UID: \"2774171b-e907-4eee-a0cd-4f8c1871ad45\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.721483 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.734765 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.760485 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.760551 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.761833 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.767114 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.770526 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4c8nc" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.772503 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.776020 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.777023 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.782075 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.793086 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-sztfr"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.794170 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.794586 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-k9jfb" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.800333 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-fbhjk" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.801291 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-sztfr"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.805259 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.818194 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6pdg\" (UniqueName: \"kubernetes.io/projected/a581380e-2a13-42a4-8a12-606c842cefbc-kube-api-access-v6pdg\") pod \"horizon-operator-controller-manager-9f4696d94-zd6mn\" (UID: \"a581380e-2a13-42a4-8a12-606c842cefbc\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.818276 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ckwb\" (UniqueName: \"kubernetes.io/projected/836c126e-917e-4b62-a2aa-7d4caac672e4-kube-api-access-7ckwb\") pod \"ironic-operator-controller-manager-7975b88857-f44nr\" (UID: \"836c126e-917e-4b62-a2aa-7d4caac672e4\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.818307 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkllm\" (UniqueName: \"kubernetes.io/projected/e0b2da1a-98f1-4008-864a-ad40956aef50-kube-api-access-rkllm\") pod \"infra-operator-controller-manager-7d857cc749-p5rtq\" (UID: \"e0b2da1a-98f1-4008-864a-ad40956aef50\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.818337 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0b2da1a-98f1-4008-864a-ad40956aef50-cert\") pod \"infra-operator-controller-manager-7d857cc749-p5rtq\" (UID: \"e0b2da1a-98f1-4008-864a-ad40956aef50\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.819509 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.819552 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.820555 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.820658 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.821059 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.825365 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-87q7c" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.825685 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-s6ckn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.826241 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.827564 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.835545 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-b7rxw" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.835770 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.870421 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ckwb\" (UniqueName: \"kubernetes.io/projected/836c126e-917e-4b62-a2aa-7d4caac672e4-kube-api-access-7ckwb\") pod \"ironic-operator-controller-manager-7975b88857-f44nr\" (UID: \"836c126e-917e-4b62-a2aa-7d4caac672e4\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.878996 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6pdg\" (UniqueName: \"kubernetes.io/projected/a581380e-2a13-42a4-8a12-606c842cefbc-kube-api-access-v6pdg\") pod \"horizon-operator-controller-manager-9f4696d94-zd6mn\" (UID: \"a581380e-2a13-42a4-8a12-606c842cefbc\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.885305 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.890132 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.892894 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-j4nb7" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.915130 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h"] Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.916513 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.918170 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.919466 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fncqh\" (UniqueName: \"kubernetes.io/projected/886c1e0e-7708-4c0a-adb2-bd5fc9b90477-kube-api-access-fncqh\") pod \"neutron-operator-controller-manager-64d7b59854-tmr6s\" (UID: \"886c1e0e-7708-4c0a-adb2-bd5fc9b90477\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.919503 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnf6j\" (UniqueName: \"kubernetes.io/projected/eeafa38f-b03a-4f83-bed3-3799872a37af-kube-api-access-wnf6j\") pod \"manila-operator-controller-manager-6d68dbc695-q5jhx\" (UID: \"eeafa38f-b03a-4f83-bed3-3799872a37af\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.919534 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkllm\" (UniqueName: \"kubernetes.io/projected/e0b2da1a-98f1-4008-864a-ad40956aef50-kube-api-access-rkllm\") pod \"infra-operator-controller-manager-7d857cc749-p5rtq\" (UID: \"e0b2da1a-98f1-4008-864a-ad40956aef50\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.919544 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-mgqgv" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.919572 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r2qv\" (UniqueName: \"kubernetes.io/projected/641c74c0-7259-43c6-8946-4da1b24a0f56-kube-api-access-4r2qv\") pod \"keystone-operator-controller-manager-5bd55b4bff-4sjbr\" (UID: \"641c74c0-7259-43c6-8946-4da1b24a0f56\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.919594 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0b2da1a-98f1-4008-864a-ad40956aef50-cert\") pod \"infra-operator-controller-manager-7d857cc749-p5rtq\" (UID: \"e0b2da1a-98f1-4008-864a-ad40956aef50\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.919613 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8zvw\" (UniqueName: \"kubernetes.io/projected/a1c0a727-272e-4ec0-be2f-b836a9c5d8d8-kube-api-access-k8zvw\") pod \"mariadb-operator-controller-manager-88c7-sztfr\" (UID: \"a1c0a727-272e-4ec0-be2f-b836a9c5d8d8\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.946837 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0b2da1a-98f1-4008-864a-ad40956aef50-cert\") pod \"infra-operator-controller-manager-7d857cc749-p5rtq\" (UID: \"e0b2da1a-98f1-4008-864a-ad40956aef50\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.954568 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.954908 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkllm\" (UniqueName: \"kubernetes.io/projected/e0b2da1a-98f1-4008-864a-ad40956aef50-kube-api-access-rkllm\") pod \"infra-operator-controller-manager-7d857cc749-p5rtq\" (UID: \"e0b2da1a-98f1-4008-864a-ad40956aef50\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.971204 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" Sep 29 19:24:51 crc kubenswrapper[4741]: I0929 19:24:51.973858 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.005323 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.006447 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.017996 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-lpjfv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.019047 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.021162 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r2qv\" (UniqueName: \"kubernetes.io/projected/641c74c0-7259-43c6-8946-4da1b24a0f56-kube-api-access-4r2qv\") pod \"keystone-operator-controller-manager-5bd55b4bff-4sjbr\" (UID: \"641c74c0-7259-43c6-8946-4da1b24a0f56\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.021199 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8v5j\" (UniqueName: \"kubernetes.io/projected/3a82b226-cd76-4abc-9699-2f907e8d7927-kube-api-access-j8v5j\") pod \"ovn-operator-controller-manager-9976ff44c-2qvtw\" (UID: \"3a82b226-cd76-4abc-9699-2f907e8d7927\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.021281 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8zvw\" (UniqueName: \"kubernetes.io/projected/a1c0a727-272e-4ec0-be2f-b836a9c5d8d8-kube-api-access-k8zvw\") pod \"mariadb-operator-controller-manager-88c7-sztfr\" (UID: \"a1c0a727-272e-4ec0-be2f-b836a9c5d8d8\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.021357 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.021378 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fncqh\" (UniqueName: \"kubernetes.io/projected/886c1e0e-7708-4c0a-adb2-bd5fc9b90477-kube-api-access-fncqh\") pod \"neutron-operator-controller-manager-64d7b59854-tmr6s\" (UID: \"886c1e0e-7708-4c0a-adb2-bd5fc9b90477\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.022033 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2cs9\" (UniqueName: \"kubernetes.io/projected/178cb996-2b77-4bb5-b9b4-58027e1d2974-kube-api-access-n2cs9\") pod \"nova-operator-controller-manager-c7c776c96-25l7k\" (UID: \"178cb996-2b77-4bb5-b9b4-58027e1d2974\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.022069 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbdrj\" (UniqueName: \"kubernetes.io/projected/5627e7c3-9301-4b01-a0c6-81957c86f73d-kube-api-access-jbdrj\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.022137 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnf6j\" (UniqueName: \"kubernetes.io/projected/eeafa38f-b03a-4f83-bed3-3799872a37af-kube-api-access-wnf6j\") pod \"manila-operator-controller-manager-6d68dbc695-q5jhx\" (UID: \"eeafa38f-b03a-4f83-bed3-3799872a37af\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.022158 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv2cp\" (UniqueName: \"kubernetes.io/projected/ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd-kube-api-access-rv2cp\") pod \"octavia-operator-controller-manager-76fcc6dc7c-qv4km\" (UID: \"ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.025962 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.030821 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.050595 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.051773 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r2qv\" (UniqueName: \"kubernetes.io/projected/641c74c0-7259-43c6-8946-4da1b24a0f56-kube-api-access-4r2qv\") pod \"keystone-operator-controller-manager-5bd55b4bff-4sjbr\" (UID: \"641c74c0-7259-43c6-8946-4da1b24a0f56\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.055348 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.057048 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fncqh\" (UniqueName: \"kubernetes.io/projected/886c1e0e-7708-4c0a-adb2-bd5fc9b90477-kube-api-access-fncqh\") pod \"neutron-operator-controller-manager-64d7b59854-tmr6s\" (UID: \"886c1e0e-7708-4c0a-adb2-bd5fc9b90477\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.057653 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8zvw\" (UniqueName: \"kubernetes.io/projected/a1c0a727-272e-4ec0-be2f-b836a9c5d8d8-kube-api-access-k8zvw\") pod \"mariadb-operator-controller-manager-88c7-sztfr\" (UID: \"a1c0a727-272e-4ec0-be2f-b836a9c5d8d8\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.067002 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-pnkmd" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.070559 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnf6j\" (UniqueName: \"kubernetes.io/projected/eeafa38f-b03a-4f83-bed3-3799872a37af-kube-api-access-wnf6j\") pod \"manila-operator-controller-manager-6d68dbc695-q5jhx\" (UID: \"eeafa38f-b03a-4f83-bed3-3799872a37af\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.095742 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.108679 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.120857 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.124002 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr9pl\" (UniqueName: \"kubernetes.io/projected/552e8661-bf43-4966-9aaf-2a2f9e514a41-kube-api-access-wr9pl\") pod \"placement-operator-controller-manager-589c58c6c-ljgcq\" (UID: \"552e8661-bf43-4966-9aaf-2a2f9e514a41\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.124048 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8v5j\" (UniqueName: \"kubernetes.io/projected/3a82b226-cd76-4abc-9699-2f907e8d7927-kube-api-access-j8v5j\") pod \"ovn-operator-controller-manager-9976ff44c-2qvtw\" (UID: \"3a82b226-cd76-4abc-9699-2f907e8d7927\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.124122 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.124150 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2cs9\" (UniqueName: \"kubernetes.io/projected/178cb996-2b77-4bb5-b9b4-58027e1d2974-kube-api-access-n2cs9\") pod \"nova-operator-controller-manager-c7c776c96-25l7k\" (UID: \"178cb996-2b77-4bb5-b9b4-58027e1d2974\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.124189 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbdrj\" (UniqueName: \"kubernetes.io/projected/5627e7c3-9301-4b01-a0c6-81957c86f73d-kube-api-access-jbdrj\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.124209 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv2cp\" (UniqueName: \"kubernetes.io/projected/ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd-kube-api-access-rv2cp\") pod \"octavia-operator-controller-manager-76fcc6dc7c-qv4km\" (UID: \"ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" Sep 29 19:24:52 crc kubenswrapper[4741]: E0929 19:24:52.124495 4741 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 29 19:24:52 crc kubenswrapper[4741]: E0929 19:24:52.124550 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert podName:5627e7c3-9301-4b01-a0c6-81957c86f73d nodeName:}" failed. No retries permitted until 2025-09-29 19:24:52.624535086 +0000 UTC m=+934.272324418 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-6jp4h" (UID: "5627e7c3-9301-4b01-a0c6-81957c86f73d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.145620 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8v5j\" (UniqueName: \"kubernetes.io/projected/3a82b226-cd76-4abc-9699-2f907e8d7927-kube-api-access-j8v5j\") pod \"ovn-operator-controller-manager-9976ff44c-2qvtw\" (UID: \"3a82b226-cd76-4abc-9699-2f907e8d7927\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.159415 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2cs9\" (UniqueName: \"kubernetes.io/projected/178cb996-2b77-4bb5-b9b4-58027e1d2974-kube-api-access-n2cs9\") pod \"nova-operator-controller-manager-c7c776c96-25l7k\" (UID: \"178cb996-2b77-4bb5-b9b4-58027e1d2974\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.159960 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv2cp\" (UniqueName: \"kubernetes.io/projected/ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd-kube-api-access-rv2cp\") pod \"octavia-operator-controller-manager-76fcc6dc7c-qv4km\" (UID: \"ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.162251 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.162680 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.163011 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbdrj\" (UniqueName: \"kubernetes.io/projected/5627e7c3-9301-4b01-a0c6-81957c86f73d-kube-api-access-jbdrj\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.164540 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.176252 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.182064 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-tsv4t" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.185591 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.186056 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.187292 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.219458 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.221683 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-7k2fm" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.222677 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.222756 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.225538 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lpmbf" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.227256 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgdbp\" (UniqueName: \"kubernetes.io/projected/3474d9f5-6b8b-4763-8271-1e024733ff25-kube-api-access-lgdbp\") pod \"swift-operator-controller-manager-bc7dc7bd9-bds8v\" (UID: \"3474d9f5-6b8b-4763-8271-1e024733ff25\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.227314 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr9pl\" (UniqueName: \"kubernetes.io/projected/552e8661-bf43-4966-9aaf-2a2f9e514a41-kube-api-access-wr9pl\") pod \"placement-operator-controller-manager-589c58c6c-ljgcq\" (UID: \"552e8661-bf43-4966-9aaf-2a2f9e514a41\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.246584 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.252488 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.252618 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr9pl\" (UniqueName: \"kubernetes.io/projected/552e8661-bf43-4966-9aaf-2a2f9e514a41-kube-api-access-wr9pl\") pod \"placement-operator-controller-manager-589c58c6c-ljgcq\" (UID: \"552e8661-bf43-4966-9aaf-2a2f9e514a41\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.261872 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.274240 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.276123 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.312569 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.329357 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x5gs\" (UniqueName: \"kubernetes.io/projected/a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e-kube-api-access-7x5gs\") pod \"telemetry-operator-controller-manager-b8d54b5d7-jrmzg\" (UID: \"a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.329461 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhhpj\" (UniqueName: \"kubernetes.io/projected/ac5bfa1f-c963-47e3-aa13-0bca249dcc78-kube-api-access-zhhpj\") pod \"watcher-operator-controller-manager-76669f99c-hgwfs\" (UID: \"ac5bfa1f-c963-47e3-aa13-0bca249dcc78\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.329498 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgdbp\" (UniqueName: \"kubernetes.io/projected/3474d9f5-6b8b-4763-8271-1e024733ff25-kube-api-access-lgdbp\") pod \"swift-operator-controller-manager-bc7dc7bd9-bds8v\" (UID: \"3474d9f5-6b8b-4763-8271-1e024733ff25\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.329519 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vzq6\" (UniqueName: \"kubernetes.io/projected/054b7215-807f-4ce0-bd9e-1cc564808ed6-kube-api-access-8vzq6\") pod \"test-operator-controller-manager-f66b554c6-6l5q9\" (UID: \"054b7215-807f-4ce0-bd9e-1cc564808ed6\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.347452 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.348558 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.355896 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.364439 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.364710 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.367170 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-v885q" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.367374 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgdbp\" (UniqueName: \"kubernetes.io/projected/3474d9f5-6b8b-4763-8271-1e024733ff25-kube-api-access-lgdbp\") pod \"swift-operator-controller-manager-bc7dc7bd9-bds8v\" (UID: \"3474d9f5-6b8b-4763-8271-1e024733ff25\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.368959 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.371419 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.372479 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.375814 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.376972 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-pb9d6" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.430271 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vzq6\" (UniqueName: \"kubernetes.io/projected/054b7215-807f-4ce0-bd9e-1cc564808ed6-kube-api-access-8vzq6\") pod \"test-operator-controller-manager-f66b554c6-6l5q9\" (UID: \"054b7215-807f-4ce0-bd9e-1cc564808ed6\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.430341 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x5gs\" (UniqueName: \"kubernetes.io/projected/a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e-kube-api-access-7x5gs\") pod \"telemetry-operator-controller-manager-b8d54b5d7-jrmzg\" (UID: \"a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.430456 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhhpj\" (UniqueName: \"kubernetes.io/projected/ac5bfa1f-c963-47e3-aa13-0bca249dcc78-kube-api-access-zhhpj\") pod \"watcher-operator-controller-manager-76669f99c-hgwfs\" (UID: \"ac5bfa1f-c963-47e3-aa13-0bca249dcc78\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.452408 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vzq6\" (UniqueName: \"kubernetes.io/projected/054b7215-807f-4ce0-bd9e-1cc564808ed6-kube-api-access-8vzq6\") pod \"test-operator-controller-manager-f66b554c6-6l5q9\" (UID: \"054b7215-807f-4ce0-bd9e-1cc564808ed6\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.458639 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x5gs\" (UniqueName: \"kubernetes.io/projected/a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e-kube-api-access-7x5gs\") pod \"telemetry-operator-controller-manager-b8d54b5d7-jrmzg\" (UID: \"a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.458961 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.459435 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhhpj\" (UniqueName: \"kubernetes.io/projected/ac5bfa1f-c963-47e3-aa13-0bca249dcc78-kube-api-access-zhhpj\") pod \"watcher-operator-controller-manager-76669f99c-hgwfs\" (UID: \"ac5bfa1f-c963-47e3-aa13-0bca249dcc78\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.475484 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" event={"ID":"00c10904-1ec2-4578-891e-2baa6ae5ceb7","Type":"ContainerStarted","Data":"a90a738bd99e0ade1cc0b6b8864eadfdf6d3adddeb151d9923688eac1269e0da"} Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.531513 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drqp8\" (UniqueName: \"kubernetes.io/projected/fe55fc47-9587-4c83-ac3b-8623ef3fe9db-kube-api-access-drqp8\") pod \"rabbitmq-cluster-operator-manager-79d8469568-ntkgq\" (UID: \"fe55fc47-9587-4c83-ac3b-8623ef3fe9db\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.531560 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv6t6\" (UniqueName: \"kubernetes.io/projected/6f1b6ef8-4bf6-4016-8333-1afad380afc5-kube-api-access-hv6t6\") pod \"openstack-operator-controller-manager-59f6b8f5c9-qwnsv\" (UID: \"6f1b6ef8-4bf6-4016-8333-1afad380afc5\") " pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.531634 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f1b6ef8-4bf6-4016-8333-1afad380afc5-cert\") pod \"openstack-operator-controller-manager-59f6b8f5c9-qwnsv\" (UID: \"6f1b6ef8-4bf6-4016-8333-1afad380afc5\") " pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.547633 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.588803 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.607912 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.632383 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drqp8\" (UniqueName: \"kubernetes.io/projected/fe55fc47-9587-4c83-ac3b-8623ef3fe9db-kube-api-access-drqp8\") pod \"rabbitmq-cluster-operator-manager-79d8469568-ntkgq\" (UID: \"fe55fc47-9587-4c83-ac3b-8623ef3fe9db\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.632437 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv6t6\" (UniqueName: \"kubernetes.io/projected/6f1b6ef8-4bf6-4016-8333-1afad380afc5-kube-api-access-hv6t6\") pod \"openstack-operator-controller-manager-59f6b8f5c9-qwnsv\" (UID: \"6f1b6ef8-4bf6-4016-8333-1afad380afc5\") " pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.632473 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.632521 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f1b6ef8-4bf6-4016-8333-1afad380afc5-cert\") pod \"openstack-operator-controller-manager-59f6b8f5c9-qwnsv\" (UID: \"6f1b6ef8-4bf6-4016-8333-1afad380afc5\") " pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: E0929 19:24:52.633248 4741 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 29 19:24:52 crc kubenswrapper[4741]: E0929 19:24:52.633284 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert podName:5627e7c3-9301-4b01-a0c6-81957c86f73d nodeName:}" failed. No retries permitted until 2025-09-29 19:24:53.63327186 +0000 UTC m=+935.281061192 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-6jp4h" (UID: "5627e7c3-9301-4b01-a0c6-81957c86f73d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.640011 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f1b6ef8-4bf6-4016-8333-1afad380afc5-cert\") pod \"openstack-operator-controller-manager-59f6b8f5c9-qwnsv\" (UID: \"6f1b6ef8-4bf6-4016-8333-1afad380afc5\") " pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.664123 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv6t6\" (UniqueName: \"kubernetes.io/projected/6f1b6ef8-4bf6-4016-8333-1afad380afc5-kube-api-access-hv6t6\") pod \"openstack-operator-controller-manager-59f6b8f5c9-qwnsv\" (UID: \"6f1b6ef8-4bf6-4016-8333-1afad380afc5\") " pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.664359 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drqp8\" (UniqueName: \"kubernetes.io/projected/fe55fc47-9587-4c83-ac3b-8623ef3fe9db-kube-api-access-drqp8\") pod \"rabbitmq-cluster-operator-manager-79d8469568-ntkgq\" (UID: \"fe55fc47-9587-4c83-ac3b-8623ef3fe9db\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.684168 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.691500 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.715203 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.741182 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.834025 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.838918 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd"] Sep 29 19:24:52 crc kubenswrapper[4741]: I0929 19:24:52.842801 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq"] Sep 29 19:24:52 crc kubenswrapper[4741]: W0929 19:24:52.858432 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5637f538_f88c_4990_8e83_b40e3217c76d.slice/crio-d582688b4fd1612e997826f312dfe1407cc625817cc6e9e9115c76c944e1824b WatchSource:0}: Error finding container d582688b4fd1612e997826f312dfe1407cc625817cc6e9e9115c76c944e1824b: Status 404 returned error can't find the container with id d582688b4fd1612e997826f312dfe1407cc625817cc6e9e9115c76c944e1824b Sep 29 19:24:52 crc kubenswrapper[4741]: W0929 19:24:52.863148 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f3daf24_fa9f_48f9_9f0a_df998fa76551.slice/crio-cb91c50793494ada14931d0736d8106a69123cac60a9f1b007c558a6db3ddee3 WatchSource:0}: Error finding container cb91c50793494ada14931d0736d8106a69123cac60a9f1b007c558a6db3ddee3: Status 404 returned error can't find the container with id cb91c50793494ada14931d0736d8106a69123cac60a9f1b007c558a6db3ddee3 Sep 29 19:24:53 crc kubenswrapper[4741]: W0929 19:24:53.065033 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod641c74c0_7259_43c6_8946_4da1b24a0f56.slice/crio-56493362c44d1885ca15087bd1278103d502531e69f7d4c81d74b9e49c4a5f9c WatchSource:0}: Error finding container 56493362c44d1885ca15087bd1278103d502531e69f7d4c81d74b9e49c4a5f9c: Status 404 returned error can't find the container with id 56493362c44d1885ca15087bd1278103d502531e69f7d4c81d74b9e49c4a5f9c Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.065812 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr"] Sep 29 19:24:53 crc kubenswrapper[4741]: W0929 19:24:53.067617 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod836c126e_917e_4b62_a2aa_7d4caac672e4.slice/crio-726c36ace6d15025e1e256acf43c37e4c3eed1447ba55be2cf5cdba0b8ab563b WatchSource:0}: Error finding container 726c36ace6d15025e1e256acf43c37e4c3eed1447ba55be2cf5cdba0b8ab563b: Status 404 returned error can't find the container with id 726c36ace6d15025e1e256acf43c37e4c3eed1447ba55be2cf5cdba0b8ab563b Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.071335 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.075554 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.079745 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-sztfr"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.083345 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.249801 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.254532 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.261636 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq"] Sep 29 19:24:53 crc kubenswrapper[4741]: W0929 19:24:53.266199 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod178cb996_2b77_4bb5_b9b4_58027e1d2974.slice/crio-34295ad3521cc1a6333276e141fb5239a8c45bc620c2bb7633bb38bb522cfe0f WatchSource:0}: Error finding container 34295ad3521cc1a6333276e141fb5239a8c45bc620c2bb7633bb38bb522cfe0f: Status 404 returned error can't find the container with id 34295ad3521cc1a6333276e141fb5239a8c45bc620c2bb7633bb38bb522cfe0f Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.268127 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw"] Sep 29 19:24:53 crc kubenswrapper[4741]: W0929 19:24:53.275244 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod552e8661_bf43_4966_9aaf_2a2f9e514a41.slice/crio-9dd52938e5b03129ea24dfc2ea6753242c69121187d14e18f2b4547225d37e12 WatchSource:0}: Error finding container 9dd52938e5b03129ea24dfc2ea6753242c69121187d14e18f2b4547225d37e12: Status 404 returned error can't find the container with id 9dd52938e5b03129ea24dfc2ea6753242c69121187d14e18f2b4547225d37e12 Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.276045 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v"] Sep 29 19:24:53 crc kubenswrapper[4741]: W0929 19:24:53.276572 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a82b226_cd76_4abc_9699_2f907e8d7927.slice/crio-d915c81e157c5533bde60b4b37a4a9d42fce194081a1d5028d9f487f7efd183b WatchSource:0}: Error finding container d915c81e157c5533bde60b4b37a4a9d42fce194081a1d5028d9f487f7efd183b: Status 404 returned error can't find the container with id d915c81e157c5533bde60b4b37a4a9d42fce194081a1d5028d9f487f7efd183b Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.278482 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j8v5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-9976ff44c-2qvtw_openstack-operators(3a82b226-cd76-4abc-9699-2f907e8d7927): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.280606 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wr9pl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-ljgcq_openstack-operators(552e8661-bf43-4966-9aaf-2a2f9e514a41): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.292635 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lgdbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-bc7dc7bd9-bds8v_openstack-operators(3474d9f5-6b8b-4763-8271-1e024733ff25): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.441768 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.460471 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.486398 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg"] Sep 29 19:24:53 crc kubenswrapper[4741]: W0929 19:24:53.496658 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda29fcbc5_3ec5_42e2_98fe_7eb8d453a76e.slice/crio-cdd94697f09a94cab77110889cbc164bdf0bc9e4d301ff71c1dd71e8b368393c WatchSource:0}: Error finding container cdd94697f09a94cab77110889cbc164bdf0bc9e4d301ff71c1dd71e8b368393c: Status 404 returned error can't find the container with id cdd94697f09a94cab77110889cbc164bdf0bc9e4d301ff71c1dd71e8b368393c Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.498444 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs"] Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.503028 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv"] Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.504667 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhhpj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-76669f99c-hgwfs_openstack-operators(ac5bfa1f-c963-47e3-aa13-0bca249dcc78): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.513918 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7x5gs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b8d54b5d7-jrmzg_openstack-operators(a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 19:24:53 crc kubenswrapper[4741]: W0929 19:24:53.519886 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod054b7215_807f_4ce0_bd9e_1cc564808ed6.slice/crio-7edecec82d728843e1f215bb77da4c33e01eb414d30420a24ad419d678ac7619 WatchSource:0}: Error finding container 7edecec82d728843e1f215bb77da4c33e01eb414d30420a24ad419d678ac7619: Status 404 returned error can't find the container with id 7edecec82d728843e1f215bb77da4c33e01eb414d30420a24ad419d678ac7619 Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.520530 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" event={"ID":"3a82b226-cd76-4abc-9699-2f907e8d7927","Type":"ContainerStarted","Data":"d915c81e157c5533bde60b4b37a4a9d42fce194081a1d5028d9f487f7efd183b"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.520843 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq"] Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.535427 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" podUID="3474d9f5-6b8b-4763-8271-1e024733ff25" Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.538096 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" event={"ID":"a1c0a727-272e-4ec0-be2f-b836a9c5d8d8","Type":"ContainerStarted","Data":"8e742836297dbc4e4ba9c9eb953c5b3f925f583ed3df43d519c84631f8b62be5"} Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.540652 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" podUID="552e8661-bf43-4966-9aaf-2a2f9e514a41" Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.542118 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" podUID="3a82b226-cd76-4abc-9699-2f907e8d7927" Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.550878 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8vzq6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-f66b554c6-6l5q9_openstack-operators(054b7215-807f-4ce0-bd9e-1cc564808ed6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.569867 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" event={"ID":"d3f19145-bf8f-4bb9-81cd-3fa20f066144","Type":"ContainerStarted","Data":"9eab0515d57834d04b80e75adc4259216d906307dde10ad39a506c2e6f9c6d88"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.573124 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" event={"ID":"552e8661-bf43-4966-9aaf-2a2f9e514a41","Type":"ContainerStarted","Data":"9dd52938e5b03129ea24dfc2ea6753242c69121187d14e18f2b4547225d37e12"} Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.574881 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" podUID="552e8661-bf43-4966-9aaf-2a2f9e514a41" Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.576748 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" event={"ID":"836c126e-917e-4b62-a2aa-7d4caac672e4","Type":"ContainerStarted","Data":"726c36ace6d15025e1e256acf43c37e4c3eed1447ba55be2cf5cdba0b8ab563b"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.580158 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" event={"ID":"ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd","Type":"ContainerStarted","Data":"555d193a867d0277565fb92db48024b5146b84d7f97808fe5c541136fa26f7d0"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.583520 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" event={"ID":"9f3daf24-fa9f-48f9-9f0a-df998fa76551","Type":"ContainerStarted","Data":"cb91c50793494ada14931d0736d8106a69123cac60a9f1b007c558a6db3ddee3"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.599458 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" event={"ID":"a581380e-2a13-42a4-8a12-606c842cefbc","Type":"ContainerStarted","Data":"fb910eac92a5fe0d5dd7c3f1fc35c58155f2cf30e4faf4a96d812a45743e65bd"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.605218 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" event={"ID":"178cb996-2b77-4bb5-b9b4-58027e1d2974","Type":"ContainerStarted","Data":"34295ad3521cc1a6333276e141fb5239a8c45bc620c2bb7633bb38bb522cfe0f"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.606596 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" event={"ID":"641c74c0-7259-43c6-8946-4da1b24a0f56","Type":"ContainerStarted","Data":"56493362c44d1885ca15087bd1278103d502531e69f7d4c81d74b9e49c4a5f9c"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.608899 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" event={"ID":"2774171b-e907-4eee-a0cd-4f8c1871ad45","Type":"ContainerStarted","Data":"626b5575d55095fe0058279d9fbf09efcbd63587619e50e80b8233183e60f5d4"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.610616 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" event={"ID":"3474d9f5-6b8b-4763-8271-1e024733ff25","Type":"ContainerStarted","Data":"6aa79f6da8ba6f0c353416e3b969a6e285835e0a81f5ac8b8e886cdc11f56af5"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.612412 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" event={"ID":"eeafa38f-b03a-4f83-bed3-3799872a37af","Type":"ContainerStarted","Data":"8ebbc953734ff39ef4c16fb1e04523fd4163afa422c66df1fef89206b15fdc21"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.624680 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" event={"ID":"e0b2da1a-98f1-4008-864a-ad40956aef50","Type":"ContainerStarted","Data":"adcf2939883918e3737c91d7c4e49ca5d4603cee94b5d01adcd3f7fe9de13178"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.630733 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" event={"ID":"886c1e0e-7708-4c0a-adb2-bd5fc9b90477","Type":"ContainerStarted","Data":"22a8510db2b3775e87b1bc4dd235d10978fc616a6d7a3c070ccb8db0c08bd561"} Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.632537 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" event={"ID":"5637f538-f88c-4990-8e83-b40e3217c76d","Type":"ContainerStarted","Data":"d582688b4fd1612e997826f312dfe1407cc625817cc6e9e9115c76c944e1824b"} Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.651664 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" podUID="3474d9f5-6b8b-4763-8271-1e024733ff25" Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.664330 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.678026 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5627e7c3-9301-4b01-a0c6-81957c86f73d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-6jp4h\" (UID: \"5627e7c3-9301-4b01-a0c6-81957c86f73d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.741758 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" podUID="a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e" Sep 29 19:24:53 crc kubenswrapper[4741]: I0929 19:24:53.827758 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:24:53 crc kubenswrapper[4741]: E0929 19:24:53.987420 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" podUID="ac5bfa1f-c963-47e3-aa13-0bca249dcc78" Sep 29 19:24:54 crc kubenswrapper[4741]: E0929 19:24:54.009484 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" podUID="054b7215-807f-4ce0-bd9e-1cc564808ed6" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.330263 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h"] Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.664024 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" event={"ID":"6f1b6ef8-4bf6-4016-8333-1afad380afc5","Type":"ContainerStarted","Data":"6a45cbe137dda90c568c8cf71836d56c8b06614e0bdf77d3bf5f809421154ac3"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.664067 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" event={"ID":"6f1b6ef8-4bf6-4016-8333-1afad380afc5","Type":"ContainerStarted","Data":"cbfe4316873f9bcaa5b82dcade2fca3a447369d06d352f58ac55cb46f957fc04"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.664076 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" event={"ID":"6f1b6ef8-4bf6-4016-8333-1afad380afc5","Type":"ContainerStarted","Data":"213eb0a2514376834aac76f77697814110d4a64798de57d07ad41bfb0c7a0be8"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.664893 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.667032 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" event={"ID":"3a82b226-cd76-4abc-9699-2f907e8d7927","Type":"ContainerStarted","Data":"b8b75cf351d5d9e980bbf8bf77c0248006b7598223d1c52b11855ad2e24a49cf"} Sep 29 19:24:54 crc kubenswrapper[4741]: E0929 19:24:54.668365 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" podUID="3a82b226-cd76-4abc-9699-2f907e8d7927" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.670043 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" event={"ID":"ac5bfa1f-c963-47e3-aa13-0bca249dcc78","Type":"ContainerStarted","Data":"bec8778b0b3223341e68c2ba0321fced173ef54e5bf441f2c18d69c95865ff42"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.670070 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" event={"ID":"ac5bfa1f-c963-47e3-aa13-0bca249dcc78","Type":"ContainerStarted","Data":"f09d4b687f9b81ef8d4cf936d4dea34cf5ae611bcce83b53305b1cbda6410df5"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.674468 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" event={"ID":"a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e","Type":"ContainerStarted","Data":"c9f8e68f3d1ca5505fb5c9a02a013d14e5759b078b4130ce352f91bb629f438f"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.674495 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" event={"ID":"a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e","Type":"ContainerStarted","Data":"cdd94697f09a94cab77110889cbc164bdf0bc9e4d301ff71c1dd71e8b368393c"} Sep 29 19:24:54 crc kubenswrapper[4741]: E0929 19:24:54.677333 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" podUID="a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e" Sep 29 19:24:54 crc kubenswrapper[4741]: E0929 19:24:54.677449 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" podUID="ac5bfa1f-c963-47e3-aa13-0bca249dcc78" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.678012 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" event={"ID":"3474d9f5-6b8b-4763-8271-1e024733ff25","Type":"ContainerStarted","Data":"378a1ac5d6a9946fd553712567b95b2d87d14765d9faf3a3b62438a4d7908a43"} Sep 29 19:24:54 crc kubenswrapper[4741]: E0929 19:24:54.679113 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" podUID="3474d9f5-6b8b-4763-8271-1e024733ff25" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.694868 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" event={"ID":"fe55fc47-9587-4c83-ac3b-8623ef3fe9db","Type":"ContainerStarted","Data":"a64c0ee187683b7e5a990221b1933ece106dc53b6af0d7e42ca9dffed908608c"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.718629 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" podStartSLOduration=2.718607664 podStartE2EDuration="2.718607664s" podCreationTimestamp="2025-09-29 19:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:24:54.706475509 +0000 UTC m=+936.354264841" watchObservedRunningTime="2025-09-29 19:24:54.718607664 +0000 UTC m=+936.366396996" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.727079 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" event={"ID":"054b7215-807f-4ce0-bd9e-1cc564808ed6","Type":"ContainerStarted","Data":"2c0fa52756805342b332f6226be57a2b1e448c3c6df2ae9aef2d6a70e7c598e2"} Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.727123 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" event={"ID":"054b7215-807f-4ce0-bd9e-1cc564808ed6","Type":"ContainerStarted","Data":"7edecec82d728843e1f215bb77da4c33e01eb414d30420a24ad419d678ac7619"} Sep 29 19:24:54 crc kubenswrapper[4741]: E0929 19:24:54.728730 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" podUID="054b7215-807f-4ce0-bd9e-1cc564808ed6" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.734761 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" event={"ID":"552e8661-bf43-4966-9aaf-2a2f9e514a41","Type":"ContainerStarted","Data":"cf86777973a99d469be677decade7fec816aafd3af5376521c19a2f4ca3cfe00"} Sep 29 19:24:54 crc kubenswrapper[4741]: E0929 19:24:54.751404 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" podUID="552e8661-bf43-4966-9aaf-2a2f9e514a41" Sep 29 19:24:54 crc kubenswrapper[4741]: I0929 19:24:54.759645 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" event={"ID":"5627e7c3-9301-4b01-a0c6-81957c86f73d","Type":"ContainerStarted","Data":"1ada51f3c96425a88ac266eab306616f7ee06e8943a1624d7628e13c8de78f61"} Sep 29 19:24:55 crc kubenswrapper[4741]: E0929 19:24:55.787419 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" podUID="a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e" Sep 29 19:24:55 crc kubenswrapper[4741]: E0929 19:24:55.787720 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" podUID="3a82b226-cd76-4abc-9699-2f907e8d7927" Sep 29 19:24:55 crc kubenswrapper[4741]: E0929 19:24:55.787793 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" podUID="552e8661-bf43-4966-9aaf-2a2f9e514a41" Sep 29 19:24:55 crc kubenswrapper[4741]: E0929 19:24:55.787898 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" podUID="054b7215-807f-4ce0-bd9e-1cc564808ed6" Sep 29 19:24:55 crc kubenswrapper[4741]: E0929 19:24:55.787967 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" podUID="3474d9f5-6b8b-4763-8271-1e024733ff25" Sep 29 19:24:55 crc kubenswrapper[4741]: E0929 19:24:55.788139 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" podUID="ac5bfa1f-c963-47e3-aa13-0bca249dcc78" Sep 29 19:25:00 crc kubenswrapper[4741]: E0929 19:25:00.139001 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice/crio-a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88\": RecentStats: unable to find data in memory cache]" Sep 29 19:25:01 crc kubenswrapper[4741]: I0929 19:25:01.738889 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:25:01 crc kubenswrapper[4741]: I0929 19:25:01.739133 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:25:02 crc kubenswrapper[4741]: I0929 19:25:02.720998 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-59f6b8f5c9-qwnsv" Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.861986 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" event={"ID":"eeafa38f-b03a-4f83-bed3-3799872a37af","Type":"ContainerStarted","Data":"49dae33b827dca80de66b8fed70d0347cc8f98371049b12258a7d09883fbb2f7"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.866763 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" event={"ID":"886c1e0e-7708-4c0a-adb2-bd5fc9b90477","Type":"ContainerStarted","Data":"a55b5193d939af9c97835bec40621d36284c2e3f0ed62ea46584e1b1c1564c76"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.872176 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" event={"ID":"a1c0a727-272e-4ec0-be2f-b836a9c5d8d8","Type":"ContainerStarted","Data":"401e9b0a68f61f83e835c2bdd691db5854bcbea37a69eefbea7afbb5367e2e1f"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.876440 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" event={"ID":"5627e7c3-9301-4b01-a0c6-81957c86f73d","Type":"ContainerStarted","Data":"68400ee4739f54428a904abea218ad1e1b33fe7f65ae95a4e9f161ce8c1c6255"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.892093 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" event={"ID":"a581380e-2a13-42a4-8a12-606c842cefbc","Type":"ContainerStarted","Data":"7bfa1b6c2be3a03125157e6e156e82d20e8e75233f6bc37a7d5ef2789049fcbc"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.898783 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" event={"ID":"836c126e-917e-4b62-a2aa-7d4caac672e4","Type":"ContainerStarted","Data":"7722e9357b10cfbf50c7014b13bbacde2a32d245f0fd4986a5535d95ff2335d9"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.903807 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" event={"ID":"ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd","Type":"ContainerStarted","Data":"08f08ca2fbd8d32b256b94a42868faa134e86f7cafaaee49f2e052383778fff4"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.908222 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" event={"ID":"178cb996-2b77-4bb5-b9b4-58027e1d2974","Type":"ContainerStarted","Data":"ec865b8650ca75f625379ec9f28c707619a7fc251dd7f178203e789356fd95bd"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.919860 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" event={"ID":"2774171b-e907-4eee-a0cd-4f8c1871ad45","Type":"ContainerStarted","Data":"f324082dd324611eed60fcac26ebcd6d3b397f708d66f37036941e77a757df2f"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.930971 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" event={"ID":"00c10904-1ec2-4578-891e-2baa6ae5ceb7","Type":"ContainerStarted","Data":"0f631bb4e44d1528f73260d2ec3357d79dd7bd2f70eef688fe244ad6968be46c"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.938874 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" event={"ID":"d3f19145-bf8f-4bb9-81cd-3fa20f066144","Type":"ContainerStarted","Data":"2f97955ed19760462514da1992d32a9b89fd57959d29dd6155c36ebd8f11529c"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.941535 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" event={"ID":"fe55fc47-9587-4c83-ac3b-8623ef3fe9db","Type":"ContainerStarted","Data":"6b2c4cd291c6653735d37e8ba65baa9b4dfae63b737ebd61f0ff8f492896b6d7"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.943781 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" event={"ID":"9f3daf24-fa9f-48f9-9f0a-df998fa76551","Type":"ContainerStarted","Data":"f4a3ed7d27972ecae9ab2c601e4532b92bce7ccbc71f43cd8354b3b26d6d213f"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.948575 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" event={"ID":"5637f538-f88c-4990-8e83-b40e3217c76d","Type":"ContainerStarted","Data":"1352ed294c957c3bc8c79679c3772c30c2a7eedcb3969d56b0c285780461a070"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.949864 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" event={"ID":"641c74c0-7259-43c6-8946-4da1b24a0f56","Type":"ContainerStarted","Data":"2e6285159a108c3d5bb9e113e6d03a978e000510d9f791c69b4387ccec6345c1"} Sep 29 19:25:06 crc kubenswrapper[4741]: I0929 19:25:06.967648 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-ntkgq" podStartSLOduration=2.4188202 podStartE2EDuration="14.967631049s" podCreationTimestamp="2025-09-29 19:24:52 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.519180872 +0000 UTC m=+935.166970204" lastFinishedPulling="2025-09-29 19:25:06.067991711 +0000 UTC m=+947.715781053" observedRunningTime="2025-09-29 19:25:06.962264629 +0000 UTC m=+948.610053961" watchObservedRunningTime="2025-09-29 19:25:06.967631049 +0000 UTC m=+948.615420381" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.959469 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" event={"ID":"00c10904-1ec2-4578-891e-2baa6ae5ceb7","Type":"ContainerStarted","Data":"cbd6cd33d56bff7b098602a3de904232dbdde3a8e279b56878834570d3107b9d"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.960735 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.962333 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" event={"ID":"d3f19145-bf8f-4bb9-81cd-3fa20f066144","Type":"ContainerStarted","Data":"21bdd89d95e1680d2b91754e99ac6dca2935302f1adcbfc7e7a2f08cd3397cbe"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.962838 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.964583 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" event={"ID":"eeafa38f-b03a-4f83-bed3-3799872a37af","Type":"ContainerStarted","Data":"4fcc67e80ece270fab9412ea26f27d6a781c4808bcd5913c10afb85d97f83348"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.965046 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.966712 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" event={"ID":"e0b2da1a-98f1-4008-864a-ad40956aef50","Type":"ContainerStarted","Data":"8c0d8c867408320bf82971cf5f58f93d1aaca736f5166cfff063038dda4ee1ea"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.966740 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" event={"ID":"e0b2da1a-98f1-4008-864a-ad40956aef50","Type":"ContainerStarted","Data":"619937c1da3536385651beafc02b821ac628c3e601d9748329e31f9052bd8410"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.967231 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.969734 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" event={"ID":"9f3daf24-fa9f-48f9-9f0a-df998fa76551","Type":"ContainerStarted","Data":"9d90334dd2d3a102478c2f246d9b3ba62a01f55c3bed1d4d6ec43395f959717e"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.970303 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.972086 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" event={"ID":"2774171b-e907-4eee-a0cd-4f8c1871ad45","Type":"ContainerStarted","Data":"dd9b0ac808d5a2a85310d10883deae5ed25fd22c91dd81a835b973c8b9390e41"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.972279 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.975783 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" event={"ID":"641c74c0-7259-43c6-8946-4da1b24a0f56","Type":"ContainerStarted","Data":"d59bbe87c482bff562c34cc6a2e78214bc4324cf26120688ad0056faf7cf7e25"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.975833 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.977744 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" event={"ID":"836c126e-917e-4b62-a2aa-7d4caac672e4","Type":"ContainerStarted","Data":"ab9161005e291289f4b4b52d6cd2d5990be8088049709e7fc08818f22cfcfa2b"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.977829 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.979682 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" event={"ID":"5637f538-f88c-4990-8e83-b40e3217c76d","Type":"ContainerStarted","Data":"4f3d20b2f16c23146a131d6b9b98941da7d8178aaf98973fb864f7ac9d3d6b7f"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.979746 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.981846 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" event={"ID":"ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd","Type":"ContainerStarted","Data":"cefef8affb5f540285b20da8ae4631f964ffa21a2489e7ddc6498749ddfbcb2d"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.982291 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.991717 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" event={"ID":"178cb996-2b77-4bb5-b9b4-58027e1d2974","Type":"ContainerStarted","Data":"9f42634dd217e372e1dcea0ddc65863aaa85d2c2ba902b58c84d85f370e94f11"} Sep 29 19:25:07 crc kubenswrapper[4741]: I0929 19:25:07.991925 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.001969 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" event={"ID":"a1c0a727-272e-4ec0-be2f-b836a9c5d8d8","Type":"ContainerStarted","Data":"0dab7a1fa94b17ebfeba8e8ff269ad374b53f2689cf46077508d1751b7318875"} Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.002622 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.005032 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" event={"ID":"a581380e-2a13-42a4-8a12-606c842cefbc","Type":"ContainerStarted","Data":"16ee6f9221850d67b55827382aacd3cef1161a3b8a546f0952e1090751bc7f58"} Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.005469 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.011134 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" event={"ID":"886c1e0e-7708-4c0a-adb2-bd5fc9b90477","Type":"ContainerStarted","Data":"22369be215a8c834ba36ddc690cbfc9a5fb3be57594ac7c0b8d1b9c09b33f0f2"} Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.011808 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.014108 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" podStartSLOduration=4.48267199 podStartE2EDuration="17.014098587s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.465140197 +0000 UTC m=+935.112929529" lastFinishedPulling="2025-09-29 19:25:05.996566794 +0000 UTC m=+947.644356126" observedRunningTime="2025-09-29 19:25:08.012581609 +0000 UTC m=+949.660370941" watchObservedRunningTime="2025-09-29 19:25:08.014098587 +0000 UTC m=+949.661887919" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.015906 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" podStartSLOduration=3.313224772 podStartE2EDuration="17.015899675s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:52.285819345 +0000 UTC m=+933.933608677" lastFinishedPulling="2025-09-29 19:25:05.988494248 +0000 UTC m=+947.636283580" observedRunningTime="2025-09-29 19:25:07.986806401 +0000 UTC m=+949.634595733" watchObservedRunningTime="2025-09-29 19:25:08.015899675 +0000 UTC m=+949.663689007" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.040112 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" event={"ID":"5627e7c3-9301-4b01-a0c6-81957c86f73d","Type":"ContainerStarted","Data":"6d6ae9a19c38b6ed7aa3f5ff699a12965027f7535c0c17d44dfec80c7bc6561f"} Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.040157 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.045949 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" podStartSLOduration=3.921728079 podStartE2EDuration="17.045936037s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:52.86489421 +0000 UTC m=+934.512683542" lastFinishedPulling="2025-09-29 19:25:05.989102168 +0000 UTC m=+947.636891500" observedRunningTime="2025-09-29 19:25:08.040656 +0000 UTC m=+949.688445332" watchObservedRunningTime="2025-09-29 19:25:08.045936037 +0000 UTC m=+949.693725369" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.062357 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" podStartSLOduration=3.744176607 podStartE2EDuration="17.062341338s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:52.716878914 +0000 UTC m=+934.364668246" lastFinishedPulling="2025-09-29 19:25:06.035043635 +0000 UTC m=+947.682832977" observedRunningTime="2025-09-29 19:25:08.061323136 +0000 UTC m=+949.709112468" watchObservedRunningTime="2025-09-29 19:25:08.062341338 +0000 UTC m=+949.710130670" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.081250 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" podStartSLOduration=4.173171009 podStartE2EDuration="17.081237337s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.069592486 +0000 UTC m=+934.717381818" lastFinishedPulling="2025-09-29 19:25:05.977658814 +0000 UTC m=+947.625448146" observedRunningTime="2025-09-29 19:25:08.079151981 +0000 UTC m=+949.726941313" watchObservedRunningTime="2025-09-29 19:25:08.081237337 +0000 UTC m=+949.729026669" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.105522 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" podStartSLOduration=4.194324071 podStartE2EDuration="17.105498628s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.084410786 +0000 UTC m=+934.732200108" lastFinishedPulling="2025-09-29 19:25:05.995585333 +0000 UTC m=+947.643374665" observedRunningTime="2025-09-29 19:25:08.097323088 +0000 UTC m=+949.745112430" watchObservedRunningTime="2025-09-29 19:25:08.105498628 +0000 UTC m=+949.753287960" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.158339 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" podStartSLOduration=4.043073652 podStartE2EDuration="17.158324514s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:52.862436753 +0000 UTC m=+934.510226085" lastFinishedPulling="2025-09-29 19:25:05.977687605 +0000 UTC m=+947.625476947" observedRunningTime="2025-09-29 19:25:08.15252499 +0000 UTC m=+949.800314322" watchObservedRunningTime="2025-09-29 19:25:08.158324514 +0000 UTC m=+949.806113846" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.158959 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" podStartSLOduration=4.031561266 podStartE2EDuration="17.158954474s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:52.896777962 +0000 UTC m=+934.544567294" lastFinishedPulling="2025-09-29 19:25:06.02417117 +0000 UTC m=+947.671960502" observedRunningTime="2025-09-29 19:25:08.135230951 +0000 UTC m=+949.783020283" watchObservedRunningTime="2025-09-29 19:25:08.158954474 +0000 UTC m=+949.806743806" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.173119 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" podStartSLOduration=3.909351918 podStartE2EDuration="17.173107203s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:52.714942522 +0000 UTC m=+934.362731854" lastFinishedPulling="2025-09-29 19:25:05.978697797 +0000 UTC m=+947.626487139" observedRunningTime="2025-09-29 19:25:08.172442762 +0000 UTC m=+949.820232094" watchObservedRunningTime="2025-09-29 19:25:08.173107203 +0000 UTC m=+949.820896535" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.203396 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" podStartSLOduration=4.278361096 podStartE2EDuration="17.203370153s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.069826943 +0000 UTC m=+934.717616285" lastFinishedPulling="2025-09-29 19:25:05.99483599 +0000 UTC m=+947.642625342" observedRunningTime="2025-09-29 19:25:08.199456109 +0000 UTC m=+949.847245441" watchObservedRunningTime="2025-09-29 19:25:08.203370153 +0000 UTC m=+949.851159475" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.237445 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" podStartSLOduration=4.463629876 podStartE2EDuration="17.237430644s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.272143363 +0000 UTC m=+934.919932695" lastFinishedPulling="2025-09-29 19:25:06.045944121 +0000 UTC m=+947.693733463" observedRunningTime="2025-09-29 19:25:08.231983242 +0000 UTC m=+949.879772574" watchObservedRunningTime="2025-09-29 19:25:08.237430644 +0000 UTC m=+949.885219976" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.256200 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" podStartSLOduration=4.353371477 podStartE2EDuration="17.256185089s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.076283068 +0000 UTC m=+934.724072400" lastFinishedPulling="2025-09-29 19:25:05.97909668 +0000 UTC m=+947.626886012" observedRunningTime="2025-09-29 19:25:08.252406169 +0000 UTC m=+949.900195501" watchObservedRunningTime="2025-09-29 19:25:08.256185089 +0000 UTC m=+949.903974421" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.267402 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" podStartSLOduration=4.343263926 podStartE2EDuration="17.267377924s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.064908337 +0000 UTC m=+934.712697669" lastFinishedPulling="2025-09-29 19:25:05.989022335 +0000 UTC m=+947.636811667" observedRunningTime="2025-09-29 19:25:08.266580679 +0000 UTC m=+949.914370011" watchObservedRunningTime="2025-09-29 19:25:08.267377924 +0000 UTC m=+949.915167256" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.290779 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" podStartSLOduration=5.660550648 podStartE2EDuration="17.290764247s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:54.348396326 +0000 UTC m=+935.996185658" lastFinishedPulling="2025-09-29 19:25:05.978609925 +0000 UTC m=+947.626399257" observedRunningTime="2025-09-29 19:25:08.287351758 +0000 UTC m=+949.935141090" watchObservedRunningTime="2025-09-29 19:25:08.290764247 +0000 UTC m=+949.938553569" Sep 29 19:25:08 crc kubenswrapper[4741]: I0929 19:25:08.323486 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" podStartSLOduration=4.596267735 podStartE2EDuration="17.323467444s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.261301629 +0000 UTC m=+934.909090961" lastFinishedPulling="2025-09-29 19:25:05.988501338 +0000 UTC m=+947.636290670" observedRunningTime="2025-09-29 19:25:08.300441873 +0000 UTC m=+949.948231205" watchObservedRunningTime="2025-09-29 19:25:08.323467444 +0000 UTC m=+949.971256846" Sep 29 19:25:10 crc kubenswrapper[4741]: I0929 19:25:10.055093 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" event={"ID":"3a82b226-cd76-4abc-9699-2f907e8d7927","Type":"ContainerStarted","Data":"d5ed6768374585bc5de22423a9d71bb0f172541ed3b39b2171d94dee9c002dbf"} Sep 29 19:25:10 crc kubenswrapper[4741]: I0929 19:25:10.056070 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" Sep 29 19:25:10 crc kubenswrapper[4741]: I0929 19:25:10.076592 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" podStartSLOduration=2.696673326 podStartE2EDuration="19.076573926s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.27833999 +0000 UTC m=+934.926129322" lastFinishedPulling="2025-09-29 19:25:09.65824059 +0000 UTC m=+951.306029922" observedRunningTime="2025-09-29 19:25:10.072826347 +0000 UTC m=+951.720615679" watchObservedRunningTime="2025-09-29 19:25:10.076573926 +0000 UTC m=+951.724363268" Sep 29 19:25:10 crc kubenswrapper[4741]: E0929 19:25:10.320545 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1fd1d_9a47_476a_8fda_ae251041b7dd.slice/crio-a6cd425f55f04d00415d8272e074558dc59a9271567b75ff2a0426b502f4bd88\": RecentStats: unable to find data in memory cache]" Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.065107 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" event={"ID":"552e8661-bf43-4966-9aaf-2a2f9e514a41","Type":"ContainerStarted","Data":"a1a41812162e1a11559b7ed6cfc9ab7e1e0c9598577ad7a1edb5699d19c98806"} Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.066207 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.083343 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" podStartSLOduration=2.637797547 podStartE2EDuration="20.083327773s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.28054351 +0000 UTC m=+934.928332842" lastFinishedPulling="2025-09-29 19:25:10.726073736 +0000 UTC m=+952.373863068" observedRunningTime="2025-09-29 19:25:11.080886585 +0000 UTC m=+952.728675927" watchObservedRunningTime="2025-09-29 19:25:11.083327773 +0000 UTC m=+952.731117105" Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.724758 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6d7tn" Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.737119 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-vnpfp" Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.778046 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-t5t9r" Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.960433 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-ntpxd" Sep 29 19:25:11 crc kubenswrapper[4741]: I0929 19:25:11.973800 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dkh26" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.037211 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-p5rtq" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.077284 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" event={"ID":"ac5bfa1f-c963-47e3-aa13-0bca249dcc78","Type":"ContainerStarted","Data":"6f0ffff4de6828e21913380fa8e5ac30044c3c1a32b81fae55b437d82375f61c"} Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.103111 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" podStartSLOduration=2.945722528 podStartE2EDuration="21.103088963s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.504359642 +0000 UTC m=+935.152148974" lastFinishedPulling="2025-09-29 19:25:11.661726077 +0000 UTC m=+953.309515409" observedRunningTime="2025-09-29 19:25:12.095890104 +0000 UTC m=+953.743679436" watchObservedRunningTime="2025-09-29 19:25:12.103088963 +0000 UTC m=+953.750878305" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.111772 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zd6mn" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.127683 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-4sjbr" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.164857 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-q5jhx" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.182973 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-f44nr" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.193496 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sztfr" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.231078 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-25l7k" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.252901 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-tmr6s" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.281182 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-qv4km" Sep 29 19:25:12 crc kubenswrapper[4741]: I0929 19:25:12.608953 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" Sep 29 19:25:13 crc kubenswrapper[4741]: I0929 19:25:13.833532 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-6jp4h" Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.104463 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" event={"ID":"3474d9f5-6b8b-4763-8271-1e024733ff25","Type":"ContainerStarted","Data":"c497aa99d8af5600954341b28bf5a0c42f576773c2f926cc46b11b4cb049d491"} Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.104719 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.106923 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" event={"ID":"054b7215-807f-4ce0-bd9e-1cc564808ed6","Type":"ContainerStarted","Data":"48048263144f552bc756522d20415dac99cc600910cae1d1b6120f38db2524e5"} Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.107200 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.111921 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" event={"ID":"a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e","Type":"ContainerStarted","Data":"aa265a8d51b4542312ef6be26c2d19c1cb4537ecde6de08732d06c721b8aa786"} Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.112737 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.125169 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" podStartSLOduration=3.486439465 podStartE2EDuration="24.12514791s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.292491779 +0000 UTC m=+934.940281111" lastFinishedPulling="2025-09-29 19:25:13.931200224 +0000 UTC m=+955.578989556" observedRunningTime="2025-09-29 19:25:15.124528342 +0000 UTC m=+956.772317754" watchObservedRunningTime="2025-09-29 19:25:15.12514791 +0000 UTC m=+956.772937262" Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.143632 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" podStartSLOduration=3.7582405530000003 podStartE2EDuration="24.143603527s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.550721804 +0000 UTC m=+935.198511136" lastFinishedPulling="2025-09-29 19:25:13.936084778 +0000 UTC m=+955.583874110" observedRunningTime="2025-09-29 19:25:15.138546716 +0000 UTC m=+956.786336068" watchObservedRunningTime="2025-09-29 19:25:15.143603527 +0000 UTC m=+956.791392909" Sep 29 19:25:15 crc kubenswrapper[4741]: I0929 19:25:15.157375 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" podStartSLOduration=3.71943822 podStartE2EDuration="24.157351153s" podCreationTimestamp="2025-09-29 19:24:51 +0000 UTC" firstStartedPulling="2025-09-29 19:24:53.513287755 +0000 UTC m=+935.161077087" lastFinishedPulling="2025-09-29 19:25:13.951200688 +0000 UTC m=+955.598990020" observedRunningTime="2025-09-29 19:25:15.152120407 +0000 UTC m=+956.799909739" watchObservedRunningTime="2025-09-29 19:25:15.157351153 +0000 UTC m=+956.805140485" Sep 29 19:25:22 crc kubenswrapper[4741]: I0929 19:25:22.316366 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-2qvtw" Sep 29 19:25:22 crc kubenswrapper[4741]: I0929 19:25:22.368064 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-ljgcq" Sep 29 19:25:22 crc kubenswrapper[4741]: I0929 19:25:22.462325 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-bds8v" Sep 29 19:25:22 crc kubenswrapper[4741]: I0929 19:25:22.550344 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6l5q9" Sep 29 19:25:22 crc kubenswrapper[4741]: I0929 19:25:22.592307 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-jrmzg" Sep 29 19:25:22 crc kubenswrapper[4741]: I0929 19:25:22.616471 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-hgwfs" Sep 29 19:25:31 crc kubenswrapper[4741]: I0929 19:25:31.738969 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:25:31 crc kubenswrapper[4741]: I0929 19:25:31.739677 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.889980 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8w55g"] Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.891836 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.895703 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.896166 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ktvqq" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.899005 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.902515 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.906411 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8w55g"] Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.945142 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-26b85"] Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.947592 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.956050 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Sep 29 19:25:40 crc kubenswrapper[4741]: I0929 19:25:40.962300 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-26b85"] Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.047951 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.048242 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mglc\" (UniqueName: \"kubernetes.io/projected/cab5bad8-da67-4471-a351-51d01c735d3a-kube-api-access-4mglc\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.048280 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-config\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.048309 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cp7s\" (UniqueName: \"kubernetes.io/projected/7350ab03-bd28-4d98-bd83-fe316663d089-kube-api-access-2cp7s\") pod \"dnsmasq-dns-675f4bcbfc-8w55g\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.048327 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7350ab03-bd28-4d98-bd83-fe316663d089-config\") pod \"dnsmasq-dns-675f4bcbfc-8w55g\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.149639 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.149693 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mglc\" (UniqueName: \"kubernetes.io/projected/cab5bad8-da67-4471-a351-51d01c735d3a-kube-api-access-4mglc\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.149712 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-config\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.149746 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cp7s\" (UniqueName: \"kubernetes.io/projected/7350ab03-bd28-4d98-bd83-fe316663d089-kube-api-access-2cp7s\") pod \"dnsmasq-dns-675f4bcbfc-8w55g\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.149765 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7350ab03-bd28-4d98-bd83-fe316663d089-config\") pod \"dnsmasq-dns-675f4bcbfc-8w55g\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.150647 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7350ab03-bd28-4d98-bd83-fe316663d089-config\") pod \"dnsmasq-dns-675f4bcbfc-8w55g\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.150686 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.151345 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-config\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.170362 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mglc\" (UniqueName: \"kubernetes.io/projected/cab5bad8-da67-4471-a351-51d01c735d3a-kube-api-access-4mglc\") pod \"dnsmasq-dns-78dd6ddcc-26b85\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.171870 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cp7s\" (UniqueName: \"kubernetes.io/projected/7350ab03-bd28-4d98-bd83-fe316663d089-kube-api-access-2cp7s\") pod \"dnsmasq-dns-675f4bcbfc-8w55g\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.224738 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.272117 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.632027 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8w55g"] Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.638055 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 19:25:41 crc kubenswrapper[4741]: I0929 19:25:41.698041 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-26b85"] Sep 29 19:25:41 crc kubenswrapper[4741]: W0929 19:25:41.704906 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcab5bad8_da67_4471_a351_51d01c735d3a.slice/crio-333093b6627e3ab9b4bbe7951fb086ef8bddf6b1b359f2e0e2af436777fb6187 WatchSource:0}: Error finding container 333093b6627e3ab9b4bbe7951fb086ef8bddf6b1b359f2e0e2af436777fb6187: Status 404 returned error can't find the container with id 333093b6627e3ab9b4bbe7951fb086ef8bddf6b1b359f2e0e2af436777fb6187 Sep 29 19:25:42 crc kubenswrapper[4741]: I0929 19:25:42.327025 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" event={"ID":"7350ab03-bd28-4d98-bd83-fe316663d089","Type":"ContainerStarted","Data":"7ee220f2527fd355fc7d1a6ae74b4c168e8f263d99a96c28d575b7b8baf9c28f"} Sep 29 19:25:42 crc kubenswrapper[4741]: I0929 19:25:42.328175 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" event={"ID":"cab5bad8-da67-4471-a351-51d01c735d3a","Type":"ContainerStarted","Data":"333093b6627e3ab9b4bbe7951fb086ef8bddf6b1b359f2e0e2af436777fb6187"} Sep 29 19:25:42 crc kubenswrapper[4741]: I0929 19:25:42.966667 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8w55g"] Sep 29 19:25:42 crc kubenswrapper[4741]: I0929 19:25:42.987859 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fwcnp"] Sep 29 19:25:42 crc kubenswrapper[4741]: I0929 19:25:42.990537 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.002627 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fwcnp"] Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.086206 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.097168 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-config\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.097291 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfvjq\" (UniqueName: \"kubernetes.io/projected/b00b1b9a-a476-418e-9b76-7c6e2692aa49-kube-api-access-cfvjq\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.200376 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.200463 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-config\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.200522 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfvjq\" (UniqueName: \"kubernetes.io/projected/b00b1b9a-a476-418e-9b76-7c6e2692aa49-kube-api-access-cfvjq\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.201307 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.201676 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-config\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.251837 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfvjq\" (UniqueName: \"kubernetes.io/projected/b00b1b9a-a476-418e-9b76-7c6e2692aa49-kube-api-access-cfvjq\") pod \"dnsmasq-dns-666b6646f7-fwcnp\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.290604 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-26b85"] Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.306169 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6dv2v"] Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.307739 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.316081 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.319508 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6dv2v"] Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.406602 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p75tg\" (UniqueName: \"kubernetes.io/projected/6f78e306-43ae-4bc3-9679-3ad4526887d3-kube-api-access-p75tg\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.406683 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-config\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.406765 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.510681 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p75tg\" (UniqueName: \"kubernetes.io/projected/6f78e306-43ae-4bc3-9679-3ad4526887d3-kube-api-access-p75tg\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.510744 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-config\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.510792 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.511623 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.517425 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-config\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.536084 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p75tg\" (UniqueName: \"kubernetes.io/projected/6f78e306-43ae-4bc3-9679-3ad4526887d3-kube-api-access-p75tg\") pod \"dnsmasq-dns-57d769cc4f-6dv2v\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.634254 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.861964 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fwcnp"] Sep 29 19:25:43 crc kubenswrapper[4741]: I0929 19:25:43.994474 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6dv2v"] Sep 29 19:25:44 crc kubenswrapper[4741]: W0929 19:25:44.000014 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f78e306_43ae_4bc3_9679_3ad4526887d3.slice/crio-e166796ee9a62a2cf75629e614dbce545817f08f8f60a074ab504b9acd5be462 WatchSource:0}: Error finding container e166796ee9a62a2cf75629e614dbce545817f08f8f60a074ab504b9acd5be462: Status 404 returned error can't find the container with id e166796ee9a62a2cf75629e614dbce545817f08f8f60a074ab504b9acd5be462 Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.137826 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.139103 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.141294 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.141569 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.141820 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-4mgnm" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.142486 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.142648 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.142801 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.144928 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.151108 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222329 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df6beb49-03ad-47ef-a9c7-3f37baa6d105-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222380 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222422 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df6beb49-03ad-47ef-a9c7-3f37baa6d105-pod-info\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222449 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222467 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222481 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222503 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222520 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222549 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q959n\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-kube-api-access-q959n\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222567 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.222581 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-server-conf\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324102 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df6beb49-03ad-47ef-a9c7-3f37baa6d105-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324160 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324181 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df6beb49-03ad-47ef-a9c7-3f37baa6d105-pod-info\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324208 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324224 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324241 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324263 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324280 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324310 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q959n\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-kube-api-access-q959n\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324329 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324348 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-server-conf\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324805 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.324889 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.325144 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.325602 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.325664 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.326124 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-server-conf\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.329900 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df6beb49-03ad-47ef-a9c7-3f37baa6d105-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.335114 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df6beb49-03ad-47ef-a9c7-3f37baa6d105-pod-info\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.338937 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.339233 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.345916 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q959n\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-kube-api-access-q959n\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.361334 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.367630 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" event={"ID":"6f78e306-43ae-4bc3-9679-3ad4526887d3","Type":"ContainerStarted","Data":"e166796ee9a62a2cf75629e614dbce545817f08f8f60a074ab504b9acd5be462"} Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.377307 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" event={"ID":"b00b1b9a-a476-418e-9b76-7c6e2692aa49","Type":"ContainerStarted","Data":"8214200f23bebadd2feb4169ba25a5dbf7893493b47b3e665af186130bd68fc2"} Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.467167 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.487221 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.488563 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.493938 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.494421 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.494533 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wwzq5" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.494670 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.494750 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.494901 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.495103 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.507948 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.632888 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633015 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxgqj\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-kube-api-access-jxgqj\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633074 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633378 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e0c02dc-69a9-4e60-b179-0e23842d10a4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633443 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633476 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633517 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633537 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633632 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633728 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e0c02dc-69a9-4e60-b179-0e23842d10a4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.633759 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735219 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e0c02dc-69a9-4e60-b179-0e23842d10a4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735274 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735314 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735356 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxgqj\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-kube-api-access-jxgqj\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735377 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735409 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e0c02dc-69a9-4e60-b179-0e23842d10a4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735426 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735463 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735477 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735507 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735823 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.735989 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.736976 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.737177 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.737280 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.743057 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e0c02dc-69a9-4e60-b179-0e23842d10a4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.743131 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e0c02dc-69a9-4e60-b179-0e23842d10a4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.744619 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.746317 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.748215 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.776254 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.785950 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxgqj\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-kube-api-access-jxgqj\") pod \"rabbitmq-cell1-server-0\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:44 crc kubenswrapper[4741]: I0929 19:25:44.835619 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.203135 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 19:25:45 crc kubenswrapper[4741]: W0929 19:25:45.227084 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf6beb49_03ad_47ef_a9c7_3f37baa6d105.slice/crio-52323e4c4ff52533356a2be38ec41b4ab9e9b44d885f039a3210f5f905046db6 WatchSource:0}: Error finding container 52323e4c4ff52533356a2be38ec41b4ab9e9b44d885f039a3210f5f905046db6: Status 404 returned error can't find the container with id 52323e4c4ff52533356a2be38ec41b4ab9e9b44d885f039a3210f5f905046db6 Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.361873 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.420014 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df6beb49-03ad-47ef-a9c7-3f37baa6d105","Type":"ContainerStarted","Data":"52323e4c4ff52533356a2be38ec41b4ab9e9b44d885f039a3210f5f905046db6"} Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.864442 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.866513 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.873362 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-bg28q" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.874188 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.877087 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.877555 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.877690 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.883768 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.924454 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958047 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-kolla-config\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958094 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958127 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958143 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc5gs\" (UniqueName: \"kubernetes.io/projected/fad684df-76f3-4651-81e8-9bb739cfbc4b-kube-api-access-wc5gs\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958512 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958586 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-default\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958680 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:45 crc kubenswrapper[4741]: I0929 19:25:45.958724 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-secrets\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061188 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061245 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-default\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061265 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061308 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061336 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-secrets\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061440 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-kolla-config\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061460 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061500 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061520 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc5gs\" (UniqueName: \"kubernetes.io/projected/fad684df-76f3-4651-81e8-9bb739cfbc4b-kube-api-access-wc5gs\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.061549 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.062214 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-default\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.062312 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.067625 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.068342 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-kolla-config\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.069068 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.073881 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-secrets\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.083059 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc5gs\" (UniqueName: \"kubernetes.io/projected/fad684df-76f3-4651-81e8-9bb739cfbc4b-kube-api-access-wc5gs\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.084354 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.095723 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.209067 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.451434 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8e0c02dc-69a9-4e60-b179-0e23842d10a4","Type":"ContainerStarted","Data":"80a92dcabd9ec3b723fea5b8cfae1457d22e984d367ebefb9b977d9fef4a156c"} Sep 29 19:25:46 crc kubenswrapper[4741]: I0929 19:25:46.787678 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 19:25:46 crc kubenswrapper[4741]: W0929 19:25:46.835776 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfad684df_76f3_4651_81e8_9bb739cfbc4b.slice/crio-2c2c11f363480f2bcbf633c4aec8f098841d5ce4c315bb29697fdb705aecd9cc WatchSource:0}: Error finding container 2c2c11f363480f2bcbf633c4aec8f098841d5ce4c315bb29697fdb705aecd9cc: Status 404 returned error can't find the container with id 2c2c11f363480f2bcbf633c4aec8f098841d5ce4c315bb29697fdb705aecd9cc Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.073658 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.076009 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.080194 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-nkn5x" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.080488 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.080830 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.080927 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.084260 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.179578 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.179623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.179672 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.179694 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.179942 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.179979 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.179997 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhrqj\" (UniqueName: \"kubernetes.io/projected/e3b64f81-4c6d-4d71-8930-109240d8ec42-kube-api-access-rhrqj\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.180064 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.180106 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283171 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283253 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283475 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283561 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283615 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283684 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.283703 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.284775 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhrqj\" (UniqueName: \"kubernetes.io/projected/e3b64f81-4c6d-4d71-8930-109240d8ec42-kube-api-access-rhrqj\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.286034 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.298052 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.298485 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.302280 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.308464 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.311932 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.312603 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.314133 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.321416 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhrqj\" (UniqueName: \"kubernetes.io/projected/e3b64f81-4c6d-4d71-8930-109240d8ec42-kube-api-access-rhrqj\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.322666 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.400960 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.468548 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fad684df-76f3-4651-81e8-9bb739cfbc4b","Type":"ContainerStarted","Data":"2c2c11f363480f2bcbf633c4aec8f098841d5ce4c315bb29697fdb705aecd9cc"} Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.477460 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.478486 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.480612 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.493883 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.494115 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.494264 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8z79w" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.594230 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-config-data\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.594328 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.594361 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kolla-config\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.594418 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.594447 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlwcm\" (UniqueName: \"kubernetes.io/projected/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kube-api-access-jlwcm\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.696136 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-config-data\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.696226 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.696248 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kolla-config\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.696290 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.696316 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlwcm\" (UniqueName: \"kubernetes.io/projected/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kube-api-access-jlwcm\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.697172 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kolla-config\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.697594 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-config-data\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.704479 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.711145 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.733080 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlwcm\" (UniqueName: \"kubernetes.io/projected/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kube-api-access-jlwcm\") pod \"memcached-0\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " pod="openstack/memcached-0" Sep 29 19:25:47 crc kubenswrapper[4741]: I0929 19:25:47.862687 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.299581 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.301366 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.309858 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-86dz7" Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.330168 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.443876 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6xk8\" (UniqueName: \"kubernetes.io/projected/84df25f5-0db8-457f-b964-3cdb332ebf02-kube-api-access-r6xk8\") pod \"kube-state-metrics-0\" (UID: \"84df25f5-0db8-457f-b964-3cdb332ebf02\") " pod="openstack/kube-state-metrics-0" Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.546049 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6xk8\" (UniqueName: \"kubernetes.io/projected/84df25f5-0db8-457f-b964-3cdb332ebf02-kube-api-access-r6xk8\") pod \"kube-state-metrics-0\" (UID: \"84df25f5-0db8-457f-b964-3cdb332ebf02\") " pod="openstack/kube-state-metrics-0" Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.571334 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6xk8\" (UniqueName: \"kubernetes.io/projected/84df25f5-0db8-457f-b964-3cdb332ebf02-kube-api-access-r6xk8\") pod \"kube-state-metrics-0\" (UID: \"84df25f5-0db8-457f-b964-3cdb332ebf02\") " pod="openstack/kube-state-metrics-0" Sep 29 19:25:49 crc kubenswrapper[4741]: I0929 19:25:49.630037 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.568690 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-68snt"] Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.570337 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.575029 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-57q72" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.575273 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.575458 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.582893 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-68snt"] Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.589594 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-q65b6"] Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.591113 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.631699 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q65b6"] Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719460 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kdjp\" (UniqueName: \"kubernetes.io/projected/865e963c-87bd-45aa-a4a7-95ae24dd0058-kube-api-access-8kdjp\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719510 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-combined-ca-bundle\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719531 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlcmc\" (UniqueName: \"kubernetes.io/projected/1e20498c-3a09-44e9-a4b1-c2dad8437611-kube-api-access-qlcmc\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719625 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-log\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719666 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run-ovn\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719737 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-log-ovn\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719796 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e20498c-3a09-44e9-a4b1-c2dad8437611-scripts\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719854 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-lib\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719881 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-ovn-controller-tls-certs\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719929 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/865e963c-87bd-45aa-a4a7-95ae24dd0058-scripts\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719955 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719974 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-run\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.719990 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-etc-ovs\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.821893 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-lib\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.821949 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-ovn-controller-tls-certs\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.822015 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/865e963c-87bd-45aa-a4a7-95ae24dd0058-scripts\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.822581 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-lib\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823035 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823073 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-run\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823099 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-etc-ovs\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823161 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kdjp\" (UniqueName: \"kubernetes.io/projected/865e963c-87bd-45aa-a4a7-95ae24dd0058-kube-api-access-8kdjp\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823192 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-combined-ca-bundle\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823221 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlcmc\" (UniqueName: \"kubernetes.io/projected/1e20498c-3a09-44e9-a4b1-c2dad8437611-kube-api-access-qlcmc\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823247 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-log\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823272 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run-ovn\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823318 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-log-ovn\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823350 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-etc-ovs\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823358 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e20498c-3a09-44e9-a4b1-c2dad8437611-scripts\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823523 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823703 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run-ovn\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823846 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-log-ovn\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.823968 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-run\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.824102 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-log\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.826091 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/865e963c-87bd-45aa-a4a7-95ae24dd0058-scripts\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.839112 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-combined-ca-bundle\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.841952 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kdjp\" (UniqueName: \"kubernetes.io/projected/865e963c-87bd-45aa-a4a7-95ae24dd0058-kube-api-access-8kdjp\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.843949 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlcmc\" (UniqueName: \"kubernetes.io/projected/1e20498c-3a09-44e9-a4b1-c2dad8437611-kube-api-access-qlcmc\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.844567 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-ovn-controller-tls-certs\") pod \"ovn-controller-68snt\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.845726 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e20498c-3a09-44e9-a4b1-c2dad8437611-scripts\") pod \"ovn-controller-ovs-q65b6\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.886572 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt" Sep 29 19:25:53 crc kubenswrapper[4741]: I0929 19:25:53.926670 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.374903 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.377435 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.380847 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.381205 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.381367 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wkz82" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.381655 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.381846 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.385162 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468132 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468197 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-config\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468234 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhxsp\" (UniqueName: \"kubernetes.io/projected/f5597514-2fd8-4d92-b115-05b66894ea94-kube-api-access-dhxsp\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468364 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468686 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468795 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468873 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.468927 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572570 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572633 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572671 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572704 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572745 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572764 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-config\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572799 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhxsp\" (UniqueName: \"kubernetes.io/projected/f5597514-2fd8-4d92-b115-05b66894ea94-kube-api-access-dhxsp\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.572827 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.574883 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.575523 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.576531 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.577905 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-config\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.583328 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.588226 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.589657 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.593895 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.600291 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.601523 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhxsp\" (UniqueName: \"kubernetes.io/projected/f5597514-2fd8-4d92-b115-05b66894ea94-kube-api-access-dhxsp\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.604723 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.604874 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.604980 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-w75vg" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.605089 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.608629 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.613087 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.673882 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.673920 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnn2l\" (UniqueName: \"kubernetes.io/projected/69f1c2bd-c091-42e8-8810-a04726ce9032-kube-api-access-mnn2l\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.673976 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.674099 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.674196 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.674309 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.674448 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-config\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.674520 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.741558 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.775806 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-config\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.775888 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.775960 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.775986 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnn2l\" (UniqueName: \"kubernetes.io/projected/69f1c2bd-c091-42e8-8810-a04726ce9032-kube-api-access-mnn2l\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.776067 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.776117 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.776157 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.776223 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.776774 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.776970 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-config\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.776983 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.777609 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.779823 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.780191 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.780259 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.793995 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnn2l\" (UniqueName: \"kubernetes.io/projected/69f1c2bd-c091-42e8-8810-a04726ce9032-kube-api-access-mnn2l\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.797330 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " pod="openstack/ovsdbserver-sb-0" Sep 29 19:25:56 crc kubenswrapper[4741]: I0929 19:25:56.939156 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.391145 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.391752 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jxgqj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(8e0c02dc-69a9-4e60-b179-0e23842d10a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.392966 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.432309 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.432509 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q959n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(df6beb49-03ad-47ef-a9c7-3f37baa6d105): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.433852 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.593700 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" Sep 29 19:26:00 crc kubenswrapper[4741]: E0929 19:26:00.593861 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" Sep 29 19:26:01 crc kubenswrapper[4741]: I0929 19:26:01.738980 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:26:01 crc kubenswrapper[4741]: I0929 19:26:01.739075 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:26:01 crc kubenswrapper[4741]: I0929 19:26:01.739114 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:26:01 crc kubenswrapper[4741]: I0929 19:26:01.739806 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d64910008ab65af756a5c6e2b068a8f502ccd9ab9a8a5bb1eac1cdbedf684c46"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:26:01 crc kubenswrapper[4741]: I0929 19:26:01.739858 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://d64910008ab65af756a5c6e2b068a8f502ccd9ab9a8a5bb1eac1cdbedf684c46" gracePeriod=600 Sep 29 19:26:02 crc kubenswrapper[4741]: I0929 19:26:02.617733 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="d64910008ab65af756a5c6e2b068a8f502ccd9ab9a8a5bb1eac1cdbedf684c46" exitCode=0 Sep 29 19:26:02 crc kubenswrapper[4741]: I0929 19:26:02.617788 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"d64910008ab65af756a5c6e2b068a8f502ccd9ab9a8a5bb1eac1cdbedf684c46"} Sep 29 19:26:02 crc kubenswrapper[4741]: I0929 19:26:02.617822 4741 scope.go:117] "RemoveContainer" containerID="1f264b13b6ab5a53b78801880659a4525cf694d3e6348258f41be568cdac61c6" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.583408 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.583748 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4mglc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-26b85_openstack(cab5bad8-da67-4471-a351-51d01c735d3a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.584981 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.585013 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" podUID="cab5bad8-da67-4471-a351-51d01c735d3a" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.586595 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2cp7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-8w55g_openstack(7350ab03-bd28-4d98-bd83-fe316663d089): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.587735 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" podUID="7350ab03-bd28-4d98-bd83-fe316663d089" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.611588 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.611724 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p75tg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-6dv2v_openstack(6f78e306-43ae-4bc3-9679-3ad4526887d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.612997 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.638673 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.638969 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.639090 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cfvjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-fwcnp_openstack(b00b1b9a-a476-418e-9b76-7c6e2692aa49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:04 crc kubenswrapper[4741]: E0929 19:26:04.641503 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.005691 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.106453 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7350ab03-bd28-4d98-bd83-fe316663d089-config\") pod \"7350ab03-bd28-4d98-bd83-fe316663d089\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.106860 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cp7s\" (UniqueName: \"kubernetes.io/projected/7350ab03-bd28-4d98-bd83-fe316663d089-kube-api-access-2cp7s\") pod \"7350ab03-bd28-4d98-bd83-fe316663d089\" (UID: \"7350ab03-bd28-4d98-bd83-fe316663d089\") " Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.107237 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7350ab03-bd28-4d98-bd83-fe316663d089-config" (OuterVolumeSpecName: "config") pod "7350ab03-bd28-4d98-bd83-fe316663d089" (UID: "7350ab03-bd28-4d98-bd83-fe316663d089"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.107366 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7350ab03-bd28-4d98-bd83-fe316663d089-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.114795 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7350ab03-bd28-4d98-bd83-fe316663d089-kube-api-access-2cp7s" (OuterVolumeSpecName: "kube-api-access-2cp7s") pod "7350ab03-bd28-4d98-bd83-fe316663d089" (UID: "7350ab03-bd28-4d98-bd83-fe316663d089"). InnerVolumeSpecName "kube-api-access-2cp7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.172633 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.180262 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.187570 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.209466 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-config\") pod \"cab5bad8-da67-4471-a351-51d01c735d3a\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.209509 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mglc\" (UniqueName: \"kubernetes.io/projected/cab5bad8-da67-4471-a351-51d01c735d3a-kube-api-access-4mglc\") pod \"cab5bad8-da67-4471-a351-51d01c735d3a\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.209534 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-dns-svc\") pod \"cab5bad8-da67-4471-a351-51d01c735d3a\" (UID: \"cab5bad8-da67-4471-a351-51d01c735d3a\") " Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.209754 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cp7s\" (UniqueName: \"kubernetes.io/projected/7350ab03-bd28-4d98-bd83-fe316663d089-kube-api-access-2cp7s\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.210374 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cab5bad8-da67-4471-a351-51d01c735d3a" (UID: "cab5bad8-da67-4471-a351-51d01c735d3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.210604 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-config" (OuterVolumeSpecName: "config") pod "cab5bad8-da67-4471-a351-51d01c735d3a" (UID: "cab5bad8-da67-4471-a351-51d01c735d3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.213262 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab5bad8-da67-4471-a351-51d01c735d3a-kube-api-access-4mglc" (OuterVolumeSpecName: "kube-api-access-4mglc") pod "cab5bad8-da67-4471-a351-51d01c735d3a" (UID: "cab5bad8-da67-4471-a351-51d01c735d3a"). InnerVolumeSpecName "kube-api-access-4mglc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.310805 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.310834 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mglc\" (UniqueName: \"kubernetes.io/projected/cab5bad8-da67-4471-a351-51d01c735d3a-kube-api-access-4mglc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.310846 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5bad8-da67-4471-a351-51d01c735d3a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.324962 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-68snt"] Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.328375 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 19:26:05 crc kubenswrapper[4741]: W0929 19:26:05.360369 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3b64f81_4c6d_4d71_8930_109240d8ec42.slice/crio-fabcaa4ef29605769073643955da8e00621af8d671f6bf30dc21156315f6dec8 WatchSource:0}: Error finding container fabcaa4ef29605769073643955da8e00621af8d671f6bf30dc21156315f6dec8: Status 404 returned error can't find the container with id fabcaa4ef29605769073643955da8e00621af8d671f6bf30dc21156315f6dec8 Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.441934 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 19:26:05 crc kubenswrapper[4741]: W0929 19:26:05.444234 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5597514_2fd8_4d92_b115_05b66894ea94.slice/crio-89bf1a08d6d1d7b35f54b96a50f9eb7fb63f0f2b6966359c9c14f1093319b854 WatchSource:0}: Error finding container 89bf1a08d6d1d7b35f54b96a50f9eb7fb63f0f2b6966359c9c14f1093319b854: Status 404 returned error can't find the container with id 89bf1a08d6d1d7b35f54b96a50f9eb7fb63f0f2b6966359c9c14f1093319b854 Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.642004 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84df25f5-0db8-457f-b964-3cdb332ebf02","Type":"ContainerStarted","Data":"e4eb82a5251b0f6c6ed3f995f8fe5db16c66735b092c8672e8f311e3a9850700"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.644150 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3","Type":"ContainerStarted","Data":"954aa19a272d9b6aad0c25106b859c57b0a12fd3fd572bba3b9fe7fef141343b"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.645667 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" event={"ID":"7350ab03-bd28-4d98-bd83-fe316663d089","Type":"ContainerDied","Data":"7ee220f2527fd355fc7d1a6ae74b4c168e8f263d99a96c28d575b7b8baf9c28f"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.645683 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8w55g" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.647270 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" event={"ID":"cab5bad8-da67-4471-a351-51d01c735d3a","Type":"ContainerDied","Data":"333093b6627e3ab9b4bbe7951fb086ef8bddf6b1b359f2e0e2af436777fb6187"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.647330 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-26b85" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.653272 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"f1bdfb84a67cbc508e610930001920c7a19e715be8c9f4e424f1f5b91fd396f8"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.655592 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt" event={"ID":"865e963c-87bd-45aa-a4a7-95ae24dd0058","Type":"ContainerStarted","Data":"7461f39eaefa65fec16c5f7cc9b53d37b849b5d431bfffec7926feecb7ea7a52"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.658143 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e3b64f81-4c6d-4d71-8930-109240d8ec42","Type":"ContainerStarted","Data":"a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.658168 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e3b64f81-4c6d-4d71-8930-109240d8ec42","Type":"ContainerStarted","Data":"fabcaa4ef29605769073643955da8e00621af8d671f6bf30dc21156315f6dec8"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.673370 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fad684df-76f3-4651-81e8-9bb739cfbc4b","Type":"ContainerStarted","Data":"6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311"} Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.678691 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f5597514-2fd8-4d92-b115-05b66894ea94","Type":"ContainerStarted","Data":"89bf1a08d6d1d7b35f54b96a50f9eb7fb63f0f2b6966359c9c14f1093319b854"} Sep 29 19:26:05 crc kubenswrapper[4741]: E0929 19:26:05.682090 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.716764 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-26b85"] Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.726520 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-26b85"] Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.789951 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8w55g"] Sep 29 19:26:05 crc kubenswrapper[4741]: I0929 19:26:05.794996 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8w55g"] Sep 29 19:26:06 crc kubenswrapper[4741]: I0929 19:26:06.185541 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q65b6"] Sep 29 19:26:06 crc kubenswrapper[4741]: W0929 19:26:06.198110 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e20498c_3a09_44e9_a4b1_c2dad8437611.slice/crio-c62f162f41d0b509b84e2fd845f6fdbce9fe539f3c55c6229bc63980d43622ab WatchSource:0}: Error finding container c62f162f41d0b509b84e2fd845f6fdbce9fe539f3c55c6229bc63980d43622ab: Status 404 returned error can't find the container with id c62f162f41d0b509b84e2fd845f6fdbce9fe539f3c55c6229bc63980d43622ab Sep 29 19:26:06 crc kubenswrapper[4741]: I0929 19:26:06.492577 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 19:26:06 crc kubenswrapper[4741]: I0929 19:26:06.686181 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q65b6" event={"ID":"1e20498c-3a09-44e9-a4b1-c2dad8437611","Type":"ContainerStarted","Data":"c62f162f41d0b509b84e2fd845f6fdbce9fe539f3c55c6229bc63980d43622ab"} Sep 29 19:26:07 crc kubenswrapper[4741]: I0929 19:26:07.099497 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7350ab03-bd28-4d98-bd83-fe316663d089" path="/var/lib/kubelet/pods/7350ab03-bd28-4d98-bd83-fe316663d089/volumes" Sep 29 19:26:07 crc kubenswrapper[4741]: I0929 19:26:07.099971 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab5bad8-da67-4471-a351-51d01c735d3a" path="/var/lib/kubelet/pods/cab5bad8-da67-4471-a351-51d01c735d3a/volumes" Sep 29 19:26:08 crc kubenswrapper[4741]: W0929 19:26:08.708666 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69f1c2bd_c091_42e8_8810_a04726ce9032.slice/crio-7301a38c44c8ebcd85b5c2b28c98a2a30bc19e665453c9b7f8eeb740f1aa5a22 WatchSource:0}: Error finding container 7301a38c44c8ebcd85b5c2b28c98a2a30bc19e665453c9b7f8eeb740f1aa5a22: Status 404 returned error can't find the container with id 7301a38c44c8ebcd85b5c2b28c98a2a30bc19e665453c9b7f8eeb740f1aa5a22 Sep 29 19:26:09 crc kubenswrapper[4741]: I0929 19:26:09.721584 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69f1c2bd-c091-42e8-8810-a04726ce9032","Type":"ContainerStarted","Data":"7301a38c44c8ebcd85b5c2b28c98a2a30bc19e665453c9b7f8eeb740f1aa5a22"} Sep 29 19:26:12 crc kubenswrapper[4741]: I0929 19:26:12.747018 4741 generic.go:334] "Generic (PLEG): container finished" podID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerID="6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311" exitCode=0 Sep 29 19:26:12 crc kubenswrapper[4741]: I0929 19:26:12.747112 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fad684df-76f3-4651-81e8-9bb739cfbc4b","Type":"ContainerDied","Data":"6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311"} Sep 29 19:26:14 crc kubenswrapper[4741]: I0929 19:26:14.763569 4741 generic.go:334] "Generic (PLEG): container finished" podID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerID="a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c" exitCode=0 Sep 29 19:26:14 crc kubenswrapper[4741]: I0929 19:26:14.763847 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e3b64f81-4c6d-4d71-8930-109240d8ec42","Type":"ContainerDied","Data":"a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c"} Sep 29 19:26:17 crc kubenswrapper[4741]: E0929 19:26:17.921587 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Sep 29 19:26:17 crc kubenswrapper[4741]: E0929 19:26:17.921957 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb8h655hbfh87h66bh5dch65dh7dh55fh569h548h585h64hcch54h687h5d8h8h675h599hbh566h568h5f5h5dbh698h598h5bh66bh5ffh5d4h5c6q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dhxsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(f5597514-2fd8-4d92-b115-05b66894ea94): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:18 crc kubenswrapper[4741]: E0929 19:26:18.307614 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Sep 29 19:26:18 crc kubenswrapper[4741]: E0929 19:26:18.307807 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n669h58dh66bh658hf8h5dfh57bh67fh6fh7bh58dh58fh59chd9h55ch5c5h685h9bh68bh566h677h5cfh54fhb4h56h68dh68bh5b4hbfh59h588h66fq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8kdjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-68snt_openstack(865e963c-87bd-45aa-a4a7-95ae24dd0058): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:26:18 crc kubenswrapper[4741]: E0929 19:26:18.309025 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-68snt" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" Sep 29 19:26:18 crc kubenswrapper[4741]: E0929 19:26:18.845629 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-68snt" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" Sep 29 19:26:19 crc kubenswrapper[4741]: E0929 19:26:19.242708 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Sep 29 19:26:19 crc kubenswrapper[4741]: E0929 19:26:19.242761 4741 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Sep 29 19:26:19 crc kubenswrapper[4741]: E0929 19:26:19.242925 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r6xk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(84df25f5-0db8-457f-b964-3cdb332ebf02): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 19:26:19 crc kubenswrapper[4741]: E0929 19:26:19.244117 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="84df25f5-0db8-457f-b964-3cdb332ebf02" Sep 29 19:26:21 crc kubenswrapper[4741]: E0929 19:26:21.306142 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="84df25f5-0db8-457f-b964-3cdb332ebf02" Sep 29 19:26:21 crc kubenswrapper[4741]: I0929 19:26:21.829443 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3","Type":"ContainerStarted","Data":"7c71b4d390feacc6365d3ebfe09f6e223b707ee503a77ad41f8f50b865956428"} Sep 29 19:26:21 crc kubenswrapper[4741]: I0929 19:26:21.830002 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Sep 29 19:26:21 crc kubenswrapper[4741]: I0929 19:26:21.835305 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e3b64f81-4c6d-4d71-8930-109240d8ec42","Type":"ContainerStarted","Data":"15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb"} Sep 29 19:26:21 crc kubenswrapper[4741]: I0929 19:26:21.842444 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8e0c02dc-69a9-4e60-b179-0e23842d10a4","Type":"ContainerStarted","Data":"7b27d2ea63000a29d7fab7a6fb5ae477f330f125488a691cbc9155958f46378e"} Sep 29 19:26:21 crc kubenswrapper[4741]: I0929 19:26:21.866471 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=21.54713267 podStartE2EDuration="34.866450581s" podCreationTimestamp="2025-09-29 19:25:47 +0000 UTC" firstStartedPulling="2025-09-29 19:26:05.19071062 +0000 UTC m=+1006.838499952" lastFinishedPulling="2025-09-29 19:26:18.510028511 +0000 UTC m=+1020.157817863" observedRunningTime="2025-09-29 19:26:21.845266408 +0000 UTC m=+1023.493055740" watchObservedRunningTime="2025-09-29 19:26:21.866450581 +0000 UTC m=+1023.514239913" Sep 29 19:26:21 crc kubenswrapper[4741]: I0929 19:26:21.874606 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=35.874587419 podStartE2EDuration="35.874587419s" podCreationTimestamp="2025-09-29 19:25:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:26:21.86895929 +0000 UTC m=+1023.516748622" watchObservedRunningTime="2025-09-29 19:26:21.874587419 +0000 UTC m=+1023.522376751" Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.851947 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69f1c2bd-c091-42e8-8810-a04726ce9032","Type":"ContainerStarted","Data":"814a325d7b6c7b6ac7fd7b9d3bac9dffa54a1007b3dabf568b1a0a21bc488445"} Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.854045 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fad684df-76f3-4651-81e8-9bb739cfbc4b","Type":"ContainerStarted","Data":"535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173"} Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.855458 4741 generic.go:334] "Generic (PLEG): container finished" podID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerID="fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81" exitCode=0 Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.855500 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" event={"ID":"b00b1b9a-a476-418e-9b76-7c6e2692aa49","Type":"ContainerDied","Data":"fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81"} Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.857928 4741 generic.go:334] "Generic (PLEG): container finished" podID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerID="47271d4927bede8c79a7cf818a442b18af36fd1d34140c9866ca66e5f484c36c" exitCode=0 Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.857990 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" event={"ID":"6f78e306-43ae-4bc3-9679-3ad4526887d3","Type":"ContainerDied","Data":"47271d4927bede8c79a7cf818a442b18af36fd1d34140c9866ca66e5f484c36c"} Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.859804 4741 generic.go:334] "Generic (PLEG): container finished" podID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerID="891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa" exitCode=0 Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.859847 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q65b6" event={"ID":"1e20498c-3a09-44e9-a4b1-c2dad8437611","Type":"ContainerDied","Data":"891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa"} Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.862057 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df6beb49-03ad-47ef-a9c7-3f37baa6d105","Type":"ContainerStarted","Data":"298a19b9e2f8e061524da9a8a4b1326a3050a055dc372e86be58a14f3e42e0e3"} Sep 29 19:26:22 crc kubenswrapper[4741]: I0929 19:26:22.873156 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.095222103 podStartE2EDuration="38.873140695s" podCreationTimestamp="2025-09-29 19:25:44 +0000 UTC" firstStartedPulling="2025-09-29 19:25:46.841043645 +0000 UTC m=+988.488832977" lastFinishedPulling="2025-09-29 19:26:04.618962247 +0000 UTC m=+1006.266751569" observedRunningTime="2025-09-29 19:26:22.869896462 +0000 UTC m=+1024.517685784" watchObservedRunningTime="2025-09-29 19:26:22.873140695 +0000 UTC m=+1024.520930027" Sep 29 19:26:23 crc kubenswrapper[4741]: E0929 19:26:23.668760 4741 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.227:59548->38.102.83.227:37241: read tcp 38.102.83.227:59548->38.102.83.227:37241: read: connection reset by peer Sep 29 19:26:24 crc kubenswrapper[4741]: E0929 19:26:24.066094 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.891695 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q65b6" event={"ID":"1e20498c-3a09-44e9-a4b1-c2dad8437611","Type":"ContainerStarted","Data":"e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f"} Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.891928 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.891944 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q65b6" event={"ID":"1e20498c-3a09-44e9-a4b1-c2dad8437611","Type":"ContainerStarted","Data":"f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14"} Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.893931 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69f1c2bd-c091-42e8-8810-a04726ce9032","Type":"ContainerStarted","Data":"b10e00d6e3d0b86eee6df15910497c173193cb10140a8e44a1e3e58aa399a2cd"} Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.896146 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" event={"ID":"b00b1b9a-a476-418e-9b76-7c6e2692aa49","Type":"ContainerStarted","Data":"641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7"} Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.896382 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.898627 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" event={"ID":"6f78e306-43ae-4bc3-9679-3ad4526887d3","Type":"ContainerStarted","Data":"96d2b5a905f059806449b941a4daaa6474579ee6d497ab007c5b1b50d529531a"} Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.898887 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.900276 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f5597514-2fd8-4d92-b115-05b66894ea94","Type":"ContainerStarted","Data":"af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188"} Sep 29 19:26:24 crc kubenswrapper[4741]: E0929 19:26:24.902093 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.914939 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-q65b6" podStartSLOduration=16.809473996 podStartE2EDuration="31.914919207s" podCreationTimestamp="2025-09-29 19:25:53 +0000 UTC" firstStartedPulling="2025-09-29 19:26:06.20070624 +0000 UTC m=+1007.848495562" lastFinishedPulling="2025-09-29 19:26:21.306151401 +0000 UTC m=+1022.953940773" observedRunningTime="2025-09-29 19:26:24.914753301 +0000 UTC m=+1026.562542643" watchObservedRunningTime="2025-09-29 19:26:24.914919207 +0000 UTC m=+1026.562708539" Sep 29 19:26:24 crc kubenswrapper[4741]: I0929 19:26:24.940744 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" podStartSLOduration=5.24270954 podStartE2EDuration="42.940716995s" podCreationTimestamp="2025-09-29 19:25:42 +0000 UTC" firstStartedPulling="2025-09-29 19:25:43.871851624 +0000 UTC m=+985.519640956" lastFinishedPulling="2025-09-29 19:26:21.569859079 +0000 UTC m=+1023.217648411" observedRunningTime="2025-09-29 19:26:24.930371087 +0000 UTC m=+1026.578160439" watchObservedRunningTime="2025-09-29 19:26:24.940716995 +0000 UTC m=+1026.588506347" Sep 29 19:26:25 crc kubenswrapper[4741]: I0929 19:26:25.016540 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" podStartSLOduration=4.478532383 podStartE2EDuration="42.01652291s" podCreationTimestamp="2025-09-29 19:25:43 +0000 UTC" firstStartedPulling="2025-09-29 19:25:44.002595173 +0000 UTC m=+985.650384505" lastFinishedPulling="2025-09-29 19:26:21.5405857 +0000 UTC m=+1023.188375032" observedRunningTime="2025-09-29 19:26:25.011599644 +0000 UTC m=+1026.659388976" watchObservedRunningTime="2025-09-29 19:26:25.01652291 +0000 UTC m=+1026.664312242" Sep 29 19:26:25 crc kubenswrapper[4741]: I0929 19:26:25.042659 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=14.900493655 podStartE2EDuration="30.042641209s" podCreationTimestamp="2025-09-29 19:25:55 +0000 UTC" firstStartedPulling="2025-09-29 19:26:08.71953218 +0000 UTC m=+1010.367321532" lastFinishedPulling="2025-09-29 19:26:23.861679754 +0000 UTC m=+1025.509469086" observedRunningTime="2025-09-29 19:26:25.038067885 +0000 UTC m=+1026.685857227" watchObservedRunningTime="2025-09-29 19:26:25.042641209 +0000 UTC m=+1026.690430541" Sep 29 19:26:25 crc kubenswrapper[4741]: I0929 19:26:25.908702 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:26:25 crc kubenswrapper[4741]: E0929 19:26:25.909896 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" Sep 29 19:26:26 crc kubenswrapper[4741]: I0929 19:26:26.209604 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Sep 29 19:26:26 crc kubenswrapper[4741]: I0929 19:26:26.209668 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Sep 29 19:26:26 crc kubenswrapper[4741]: I0929 19:26:26.254984 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Sep 29 19:26:26 crc kubenswrapper[4741]: I0929 19:26:26.939587 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Sep 29 19:26:26 crc kubenswrapper[4741]: I0929 19:26:26.939643 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Sep 29 19:26:26 crc kubenswrapper[4741]: I0929 19:26:26.960075 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Sep 29 19:26:26 crc kubenswrapper[4741]: I0929 19:26:26.983003 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.401402 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.401456 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.451227 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.508349 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-hqmd7"] Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.509902 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hqmd7" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.527664 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hqmd7"] Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.589870 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw5ml\" (UniqueName: \"kubernetes.io/projected/2a8d0551-6f96-4651-908d-3a048f691ddb-kube-api-access-tw5ml\") pod \"keystone-db-create-hqmd7\" (UID: \"2a8d0551-6f96-4651-908d-3a048f691ddb\") " pod="openstack/keystone-db-create-hqmd7" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.690921 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw5ml\" (UniqueName: \"kubernetes.io/projected/2a8d0551-6f96-4651-908d-3a048f691ddb-kube-api-access-tw5ml\") pod \"keystone-db-create-hqmd7\" (UID: \"2a8d0551-6f96-4651-908d-3a048f691ddb\") " pod="openstack/keystone-db-create-hqmd7" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.717563 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw5ml\" (UniqueName: \"kubernetes.io/projected/2a8d0551-6f96-4651-908d-3a048f691ddb-kube-api-access-tw5ml\") pod \"keystone-db-create-hqmd7\" (UID: \"2a8d0551-6f96-4651-908d-3a048f691ddb\") " pod="openstack/keystone-db-create-hqmd7" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.770111 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-fqvpl"] Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.771478 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqvpl" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.781248 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fqvpl"] Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.826659 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hqmd7" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.863819 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Sep 29 19:26:27 crc kubenswrapper[4741]: I0929 19:26:27.921612 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pf2g\" (UniqueName: \"kubernetes.io/projected/029559ea-d2e3-4a1b-87b8-60b0a65c9730-kube-api-access-8pf2g\") pod \"placement-db-create-fqvpl\" (UID: \"029559ea-d2e3-4a1b-87b8-60b0a65c9730\") " pod="openstack/placement-db-create-fqvpl" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.012133 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.015713 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.024918 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pf2g\" (UniqueName: \"kubernetes.io/projected/029559ea-d2e3-4a1b-87b8-60b0a65c9730-kube-api-access-8pf2g\") pod \"placement-db-create-fqvpl\" (UID: \"029559ea-d2e3-4a1b-87b8-60b0a65c9730\") " pod="openstack/placement-db-create-fqvpl" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.103083 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pf2g\" (UniqueName: \"kubernetes.io/projected/029559ea-d2e3-4a1b-87b8-60b0a65c9730-kube-api-access-8pf2g\") pod \"placement-db-create-fqvpl\" (UID: \"029559ea-d2e3-4a1b-87b8-60b0a65c9730\") " pod="openstack/placement-db-create-fqvpl" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.120520 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dn5tb"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.121696 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dn5tb" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.121832 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dn5tb"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.142305 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqvpl" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.234063 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfjq5\" (UniqueName: \"kubernetes.io/projected/4eae36aa-f785-4402-ae16-3d3a68050ebb-kube-api-access-qfjq5\") pod \"glance-db-create-dn5tb\" (UID: \"4eae36aa-f785-4402-ae16-3d3a68050ebb\") " pod="openstack/glance-db-create-dn5tb" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.244197 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fwcnp"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.244420 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerName="dnsmasq-dns" containerID="cri-o://641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7" gracePeriod=10 Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.279971 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-pfjg6"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.281997 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.285214 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.292116 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-pfjg6"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.335052 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.335147 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxshx\" (UniqueName: \"kubernetes.io/projected/37fe428b-b0e2-4b27-b91d-e8407e99bf81-kube-api-access-lxshx\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.335172 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-config\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.335219 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfjq5\" (UniqueName: \"kubernetes.io/projected/4eae36aa-f785-4402-ae16-3d3a68050ebb-kube-api-access-qfjq5\") pod \"glance-db-create-dn5tb\" (UID: \"4eae36aa-f785-4402-ae16-3d3a68050ebb\") " pod="openstack/glance-db-create-dn5tb" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.335237 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.357216 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfjq5\" (UniqueName: \"kubernetes.io/projected/4eae36aa-f785-4402-ae16-3d3a68050ebb-kube-api-access-qfjq5\") pod \"glance-db-create-dn5tb\" (UID: \"4eae36aa-f785-4402-ae16-3d3a68050ebb\") " pod="openstack/glance-db-create-dn5tb" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.361557 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hqmd7"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.386414 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-fkpb5"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.387495 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.389470 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.393429 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fkpb5"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436051 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436110 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436142 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c11cbdf-1417-4c8d-b3bb-150f57496287-config\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436165 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovn-rundir\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436213 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv5tn\" (UniqueName: \"kubernetes.io/projected/3c11cbdf-1417-4c8d-b3bb-150f57496287-kube-api-access-sv5tn\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436248 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxshx\" (UniqueName: \"kubernetes.io/projected/37fe428b-b0e2-4b27-b91d-e8407e99bf81-kube-api-access-lxshx\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436269 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-config\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436289 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-combined-ca-bundle\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436315 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovs-rundir\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.436330 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.437072 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.437093 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.437617 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-config\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.449548 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dn5tb" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.457049 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxshx\" (UniqueName: \"kubernetes.io/projected/37fe428b-b0e2-4b27-b91d-e8407e99bf81-kube-api-access-lxshx\") pod \"dnsmasq-dns-7f896c8c65-pfjg6\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.537174 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovs-rundir\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.537212 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.537281 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c11cbdf-1417-4c8d-b3bb-150f57496287-config\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.537307 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovn-rundir\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.537384 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv5tn\" (UniqueName: \"kubernetes.io/projected/3c11cbdf-1417-4c8d-b3bb-150f57496287-kube-api-access-sv5tn\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.537488 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-combined-ca-bundle\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.537562 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovs-rundir\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.538383 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c11cbdf-1417-4c8d-b3bb-150f57496287-config\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.538529 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovn-rundir\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.541181 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-combined-ca-bundle\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.541869 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.564415 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv5tn\" (UniqueName: \"kubernetes.io/projected/3c11cbdf-1417-4c8d-b3bb-150f57496287-kube-api-access-sv5tn\") pod \"ovn-controller-metrics-fkpb5\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.671940 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6dv2v"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.672218 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerName="dnsmasq-dns" containerID="cri-o://96d2b5a905f059806449b941a4daaa6474579ee6d497ab007c5b1b50d529531a" gracePeriod=10 Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.692551 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fqvpl"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.692867 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.700541 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tvwnx"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.702236 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.712672 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.713484 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.731279 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tvwnx"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.741359 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvpxc\" (UniqueName: \"kubernetes.io/projected/a5a18fae-a329-44dd-9bf9-75944ef94b62-kube-api-access-kvpxc\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.741488 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.741562 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-config\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.741623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.741649 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.785881 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dn5tb"] Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.842969 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.843019 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.843078 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvpxc\" (UniqueName: \"kubernetes.io/projected/a5a18fae-a329-44dd-9bf9-75944ef94b62-kube-api-access-kvpxc\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.843107 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.843154 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-config\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.844080 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-config\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.844802 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.845192 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.845480 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.845521 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: W0929 19:26:28.858028 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eae36aa_f785_4402_ae16_3d3a68050ebb.slice/crio-bf5e8a6d3d4a1f472ba723f77ec00d996861c767e8edf252351f90a06ab2c23c WatchSource:0}: Error finding container bf5e8a6d3d4a1f472ba723f77ec00d996861c767e8edf252351f90a06ab2c23c: Status 404 returned error can't find the container with id bf5e8a6d3d4a1f472ba723f77ec00d996861c767e8edf252351f90a06ab2c23c Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.896427 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvpxc\" (UniqueName: \"kubernetes.io/projected/a5a18fae-a329-44dd-9bf9-75944ef94b62-kube-api-access-kvpxc\") pod \"dnsmasq-dns-86db49b7ff-tvwnx\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.945869 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-dns-svc\") pod \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.945929 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-config\") pod \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.945978 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfvjq\" (UniqueName: \"kubernetes.io/projected/b00b1b9a-a476-418e-9b76-7c6e2692aa49-kube-api-access-cfvjq\") pod \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\" (UID: \"b00b1b9a-a476-418e-9b76-7c6e2692aa49\") " Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.963634 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00b1b9a-a476-418e-9b76-7c6e2692aa49-kube-api-access-cfvjq" (OuterVolumeSpecName: "kube-api-access-cfvjq") pod "b00b1b9a-a476-418e-9b76-7c6e2692aa49" (UID: "b00b1b9a-a476-418e-9b76-7c6e2692aa49"). InnerVolumeSpecName "kube-api-access-cfvjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.969611 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dn5tb" event={"ID":"4eae36aa-f785-4402-ae16-3d3a68050ebb","Type":"ContainerStarted","Data":"bf5e8a6d3d4a1f472ba723f77ec00d996861c767e8edf252351f90a06ab2c23c"} Sep 29 19:26:28 crc kubenswrapper[4741]: I0929 19:26:28.999210 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b00b1b9a-a476-418e-9b76-7c6e2692aa49" (UID: "b00b1b9a-a476-418e-9b76-7c6e2692aa49"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.015314 4741 generic.go:334] "Generic (PLEG): container finished" podID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerID="641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7" exitCode=0 Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.015392 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" event={"ID":"b00b1b9a-a476-418e-9b76-7c6e2692aa49","Type":"ContainerDied","Data":"641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7"} Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.015466 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" event={"ID":"b00b1b9a-a476-418e-9b76-7c6e2692aa49","Type":"ContainerDied","Data":"8214200f23bebadd2feb4169ba25a5dbf7893493b47b3e665af186130bd68fc2"} Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.015489 4741 scope.go:117] "RemoveContainer" containerID="641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.015577 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fwcnp" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.047100 4741 generic.go:334] "Generic (PLEG): container finished" podID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerID="96d2b5a905f059806449b941a4daaa6474579ee6d497ab007c5b1b50d529531a" exitCode=0 Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.047198 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" event={"ID":"6f78e306-43ae-4bc3-9679-3ad4526887d3","Type":"ContainerDied","Data":"96d2b5a905f059806449b941a4daaa6474579ee6d497ab007c5b1b50d529531a"} Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.047762 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.047823 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfvjq\" (UniqueName: \"kubernetes.io/projected/b00b1b9a-a476-418e-9b76-7c6e2692aa49-kube-api-access-cfvjq\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.057432 4741 generic.go:334] "Generic (PLEG): container finished" podID="2a8d0551-6f96-4651-908d-3a048f691ddb" containerID="d131ad911bbb73579247dd956ee8718242e5a384d4d827fd0a111dfc506babf9" exitCode=0 Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.057497 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hqmd7" event={"ID":"2a8d0551-6f96-4651-908d-3a048f691ddb","Type":"ContainerDied","Data":"d131ad911bbb73579247dd956ee8718242e5a384d4d827fd0a111dfc506babf9"} Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.057523 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hqmd7" event={"ID":"2a8d0551-6f96-4651-908d-3a048f691ddb","Type":"ContainerStarted","Data":"06d99d9b9a9ac9469190e4beab5a58033345d06c9af75868e9b43362404937c5"} Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.059676 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fqvpl" event={"ID":"029559ea-d2e3-4a1b-87b8-60b0a65c9730","Type":"ContainerStarted","Data":"912dd288865183d61a5c6614c7a6a6a915eb8733a27f506cb408d03a5ff1d3a8"} Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.071528 4741 scope.go:117] "RemoveContainer" containerID="fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.079539 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-config" (OuterVolumeSpecName: "config") pod "b00b1b9a-a476-418e-9b76-7c6e2692aa49" (UID: "b00b1b9a-a476-418e-9b76-7c6e2692aa49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.165595 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.177678 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b00b1b9a-a476-418e-9b76-7c6e2692aa49-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.179302 4741 scope.go:117] "RemoveContainer" containerID="641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7" Sep 29 19:26:29 crc kubenswrapper[4741]: E0929 19:26:29.180887 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7\": container with ID starting with 641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7 not found: ID does not exist" containerID="641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.180952 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7"} err="failed to get container status \"641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7\": rpc error: code = NotFound desc = could not find container \"641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7\": container with ID starting with 641f639449c98088dbda7b9057a17e951c68d861bd8d06b67561b20c967015d7 not found: ID does not exist" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.180982 4741 scope.go:117] "RemoveContainer" containerID="fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81" Sep 29 19:26:29 crc kubenswrapper[4741]: E0929 19:26:29.183265 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81\": container with ID starting with fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81 not found: ID does not exist" containerID="fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.183335 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81"} err="failed to get container status \"fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81\": rpc error: code = NotFound desc = could not find container \"fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81\": container with ID starting with fff75b5b103d7ec6d2d9b09dba01518b1c093c40a91243c83adcde70c94ded81 not found: ID does not exist" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.392829 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.406965 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fwcnp"] Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.411481 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fwcnp"] Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.484102 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p75tg\" (UniqueName: \"kubernetes.io/projected/6f78e306-43ae-4bc3-9679-3ad4526887d3-kube-api-access-p75tg\") pod \"6f78e306-43ae-4bc3-9679-3ad4526887d3\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.484150 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-dns-svc\") pod \"6f78e306-43ae-4bc3-9679-3ad4526887d3\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.484325 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-config\") pod \"6f78e306-43ae-4bc3-9679-3ad4526887d3\" (UID: \"6f78e306-43ae-4bc3-9679-3ad4526887d3\") " Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.490484 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f78e306-43ae-4bc3-9679-3ad4526887d3-kube-api-access-p75tg" (OuterVolumeSpecName: "kube-api-access-p75tg") pod "6f78e306-43ae-4bc3-9679-3ad4526887d3" (UID: "6f78e306-43ae-4bc3-9679-3ad4526887d3"). InnerVolumeSpecName "kube-api-access-p75tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.534796 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-config" (OuterVolumeSpecName: "config") pod "6f78e306-43ae-4bc3-9679-3ad4526887d3" (UID: "6f78e306-43ae-4bc3-9679-3ad4526887d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.540119 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fkpb5"] Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.587975 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.588010 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p75tg\" (UniqueName: \"kubernetes.io/projected/6f78e306-43ae-4bc3-9679-3ad4526887d3-kube-api-access-p75tg\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.596390 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f78e306-43ae-4bc3-9679-3ad4526887d3" (UID: "6f78e306-43ae-4bc3-9679-3ad4526887d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.612319 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-pfjg6"] Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.662767 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-sxlh8"] Sep 29 19:26:29 crc kubenswrapper[4741]: E0929 19:26:29.663113 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerName="init" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.663129 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerName="init" Sep 29 19:26:29 crc kubenswrapper[4741]: E0929 19:26:29.663144 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerName="dnsmasq-dns" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.663151 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerName="dnsmasq-dns" Sep 29 19:26:29 crc kubenswrapper[4741]: E0929 19:26:29.663162 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerName="dnsmasq-dns" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.663170 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerName="dnsmasq-dns" Sep 29 19:26:29 crc kubenswrapper[4741]: E0929 19:26:29.663200 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerName="init" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.663206 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerName="init" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.663346 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" containerName="dnsmasq-dns" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.663371 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" containerName="dnsmasq-dns" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.664166 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: W0929 19:26:29.672989 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37fe428b_b0e2_4b27_b91d_e8407e99bf81.slice/crio-37cea62349f673df2dfd28c44750fae50d710e059455293bab1b4ddf3206bf54 WatchSource:0}: Error finding container 37cea62349f673df2dfd28c44750fae50d710e059455293bab1b4ddf3206bf54: Status 404 returned error can't find the container with id 37cea62349f673df2dfd28c44750fae50d710e059455293bab1b4ddf3206bf54 Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.692037 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f78e306-43ae-4bc3-9679-3ad4526887d3-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.695048 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-pfjg6"] Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.714507 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sxlh8"] Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.726344 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tvwnx"] Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.797220 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.797272 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-config\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.797293 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkn9w\" (UniqueName: \"kubernetes.io/projected/07649762-f417-4bfe-877a-1526d40f2498-kube-api-access-gkn9w\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.797316 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.797344 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-dns-svc\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.898833 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.898911 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-config\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.898935 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkn9w\" (UniqueName: \"kubernetes.io/projected/07649762-f417-4bfe-877a-1526d40f2498-kube-api-access-gkn9w\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.898979 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.899010 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-dns-svc\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.899747 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.899828 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-config\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.900079 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.901502 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-dns-svc\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:29 crc kubenswrapper[4741]: I0929 19:26:29.920741 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkn9w\" (UniqueName: \"kubernetes.io/projected/07649762-f417-4bfe-877a-1526d40f2498-kube-api-access-gkn9w\") pod \"dnsmasq-dns-698758b865-sxlh8\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.004274 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.073692 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fkpb5" event={"ID":"3c11cbdf-1417-4c8d-b3bb-150f57496287","Type":"ContainerStarted","Data":"ae21046b33ac269496303f0c89312707199e95ccc69f985f150deaee38b8c665"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.073746 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fkpb5" event={"ID":"3c11cbdf-1417-4c8d-b3bb-150f57496287","Type":"ContainerStarted","Data":"2049ecd7447da69f0f6a9aaa984099c27e8611d3ad2551ceab940b97a3aae395"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.080012 4741 generic.go:334] "Generic (PLEG): container finished" podID="37fe428b-b0e2-4b27-b91d-e8407e99bf81" containerID="42257dbd9289b2096acd1e52d51a22000043236c47e957f4dc8a6e60ec5997ad" exitCode=0 Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.080113 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" event={"ID":"37fe428b-b0e2-4b27-b91d-e8407e99bf81","Type":"ContainerDied","Data":"42257dbd9289b2096acd1e52d51a22000043236c47e957f4dc8a6e60ec5997ad"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.080562 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" event={"ID":"37fe428b-b0e2-4b27-b91d-e8407e99bf81","Type":"ContainerStarted","Data":"37cea62349f673df2dfd28c44750fae50d710e059455293bab1b4ddf3206bf54"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.081881 4741 generic.go:334] "Generic (PLEG): container finished" podID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerID="d6b611a9627ae6bee6c41ccb8652ca3dc9faffd81accc380242d52cb411f881e" exitCode=0 Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.081942 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" event={"ID":"a5a18fae-a329-44dd-9bf9-75944ef94b62","Type":"ContainerDied","Data":"d6b611a9627ae6bee6c41ccb8652ca3dc9faffd81accc380242d52cb411f881e"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.081971 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" event={"ID":"a5a18fae-a329-44dd-9bf9-75944ef94b62","Type":"ContainerStarted","Data":"8e5fa4dcdef906ef42794f8e12034e9fc16b6bae9058e55a04cdd14f235b472b"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.097886 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-fkpb5" podStartSLOduration=2.097872075 podStartE2EDuration="2.097872075s" podCreationTimestamp="2025-09-29 19:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:26:30.093229228 +0000 UTC m=+1031.741018580" watchObservedRunningTime="2025-09-29 19:26:30.097872075 +0000 UTC m=+1031.745661407" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.103511 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" event={"ID":"6f78e306-43ae-4bc3-9679-3ad4526887d3","Type":"ContainerDied","Data":"e166796ee9a62a2cf75629e614dbce545817f08f8f60a074ab504b9acd5be462"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.103547 4741 scope.go:117] "RemoveContainer" containerID="96d2b5a905f059806449b941a4daaa6474579ee6d497ab007c5b1b50d529531a" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.103621 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6dv2v" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.111468 4741 generic.go:334] "Generic (PLEG): container finished" podID="029559ea-d2e3-4a1b-87b8-60b0a65c9730" containerID="64a5905d0be5104788453fc11b7eaaf36923a36ee162da53e9e8481b08c87205" exitCode=0 Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.111602 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fqvpl" event={"ID":"029559ea-d2e3-4a1b-87b8-60b0a65c9730","Type":"ContainerDied","Data":"64a5905d0be5104788453fc11b7eaaf36923a36ee162da53e9e8481b08c87205"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.127645 4741 generic.go:334] "Generic (PLEG): container finished" podID="4eae36aa-f785-4402-ae16-3d3a68050ebb" containerID="682294e2a19501673b2a62943ea271a293eba88bea8c3d9e3f37a7422b35cdf7" exitCode=0 Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.127903 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dn5tb" event={"ID":"4eae36aa-f785-4402-ae16-3d3a68050ebb","Type":"ContainerDied","Data":"682294e2a19501673b2a62943ea271a293eba88bea8c3d9e3f37a7422b35cdf7"} Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.193985 4741 scope.go:117] "RemoveContainer" containerID="47271d4927bede8c79a7cf818a442b18af36fd1d34140c9866ca66e5f484c36c" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.217306 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6dv2v"] Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.223044 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6dv2v"] Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.341614 4741 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Sep 29 19:26:30 crc kubenswrapper[4741]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/a5a18fae-a329-44dd-9bf9-75944ef94b62/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Sep 29 19:26:30 crc kubenswrapper[4741]: > podSandboxID="8e5fa4dcdef906ef42794f8e12034e9fc16b6bae9058e55a04cdd14f235b472b" Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.341775 4741 kuberuntime_manager.go:1274] "Unhandled Error" err=< Sep 29 19:26:30 crc kubenswrapper[4741]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n599h5cbh7ch5d4h66fh676hdbh546h95h88h5ffh55ch7fhch57ch687hddhc7h5fdh57dh674h56fh64ch98h9bh557h55dh646h54ch54fh5c4h597q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kvpxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-86db49b7ff-tvwnx_openstack(a5a18fae-a329-44dd-9bf9-75944ef94b62): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/a5a18fae-a329-44dd-9bf9-75944ef94b62/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Sep 29 19:26:30 crc kubenswrapper[4741]: > logger="UnhandledError" Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.342936 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/a5a18fae-a329-44dd-9bf9-75944ef94b62/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.443852 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.486930 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hqmd7" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.508315 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-dns-svc\") pod \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.508447 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-ovsdbserver-sb\") pod \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.508577 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-config\") pod \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.508648 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxshx\" (UniqueName: \"kubernetes.io/projected/37fe428b-b0e2-4b27-b91d-e8407e99bf81-kube-api-access-lxshx\") pod \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\" (UID: \"37fe428b-b0e2-4b27-b91d-e8407e99bf81\") " Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.517626 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37fe428b-b0e2-4b27-b91d-e8407e99bf81-kube-api-access-lxshx" (OuterVolumeSpecName: "kube-api-access-lxshx") pod "37fe428b-b0e2-4b27-b91d-e8407e99bf81" (UID: "37fe428b-b0e2-4b27-b91d-e8407e99bf81"). InnerVolumeSpecName "kube-api-access-lxshx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.527760 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-config" (OuterVolumeSpecName: "config") pod "37fe428b-b0e2-4b27-b91d-e8407e99bf81" (UID: "37fe428b-b0e2-4b27-b91d-e8407e99bf81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.528595 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37fe428b-b0e2-4b27-b91d-e8407e99bf81" (UID: "37fe428b-b0e2-4b27-b91d-e8407e99bf81"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.532964 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37fe428b-b0e2-4b27-b91d-e8407e99bf81" (UID: "37fe428b-b0e2-4b27-b91d-e8407e99bf81"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.610480 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw5ml\" (UniqueName: \"kubernetes.io/projected/2a8d0551-6f96-4651-908d-3a048f691ddb-kube-api-access-tw5ml\") pod \"2a8d0551-6f96-4651-908d-3a048f691ddb\" (UID: \"2a8d0551-6f96-4651-908d-3a048f691ddb\") " Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.611019 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.611042 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxshx\" (UniqueName: \"kubernetes.io/projected/37fe428b-b0e2-4b27-b91d-e8407e99bf81-kube-api-access-lxshx\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.611050 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.611059 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fe428b-b0e2-4b27-b91d-e8407e99bf81-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.615678 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a8d0551-6f96-4651-908d-3a048f691ddb-kube-api-access-tw5ml" (OuterVolumeSpecName: "kube-api-access-tw5ml") pod "2a8d0551-6f96-4651-908d-3a048f691ddb" (UID: "2a8d0551-6f96-4651-908d-3a048f691ddb"). InnerVolumeSpecName "kube-api-access-tw5ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.647807 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sxlh8"] Sep 29 19:26:30 crc kubenswrapper[4741]: W0929 19:26:30.656817 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07649762_f417_4bfe_877a_1526d40f2498.slice/crio-8d5b08d3c1d11d655c028172403fe99c4d8d41b8284f8e04d6146ea25ef4cec1 WatchSource:0}: Error finding container 8d5b08d3c1d11d655c028172403fe99c4d8d41b8284f8e04d6146ea25ef4cec1: Status 404 returned error can't find the container with id 8d5b08d3c1d11d655c028172403fe99c4d8d41b8284f8e04d6146ea25ef4cec1 Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.699916 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.700291 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37fe428b-b0e2-4b27-b91d-e8407e99bf81" containerName="init" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.700307 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="37fe428b-b0e2-4b27-b91d-e8407e99bf81" containerName="init" Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.700322 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8d0551-6f96-4651-908d-3a048f691ddb" containerName="mariadb-database-create" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.700328 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8d0551-6f96-4651-908d-3a048f691ddb" containerName="mariadb-database-create" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.700502 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="37fe428b-b0e2-4b27-b91d-e8407e99bf81" containerName="init" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.700518 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a8d0551-6f96-4651-908d-3a048f691ddb" containerName="mariadb-database-create" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.718409 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw5ml\" (UniqueName: \"kubernetes.io/projected/2a8d0551-6f96-4651-908d-3a048f691ddb-kube-api-access-tw5ml\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.727845 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.734094 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.737702 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.738023 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.738209 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.742766 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-5dvnz" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.819977 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62jq\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-kube-api-access-d62jq\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.820357 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.820416 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-lock\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.820623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.820666 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-cache\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.922157 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.922200 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-lock\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.922253 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.922276 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-cache\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.922316 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62jq\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-kube-api-access-d62jq\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.922664 4741 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.922681 4741 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 19:26:30 crc kubenswrapper[4741]: E0929 19:26:30.922715 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift podName:95b48711-844f-4e95-9f01-36bec6b9995a nodeName:}" failed. No retries permitted until 2025-09-29 19:26:31.42270174 +0000 UTC m=+1033.070491072 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift") pod "swift-storage-0" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a") : configmap "swift-ring-files" not found Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.923245 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-lock\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.923478 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.925039 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-cache\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.939917 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62jq\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-kube-api-access-d62jq\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:30 crc kubenswrapper[4741]: I0929 19:26:30.945515 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.097085 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f78e306-43ae-4bc3-9679-3ad4526887d3" path="/var/lib/kubelet/pods/6f78e306-43ae-4bc3-9679-3ad4526887d3/volumes" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.097688 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b00b1b9a-a476-418e-9b76-7c6e2692aa49" path="/var/lib/kubelet/pods/b00b1b9a-a476-418e-9b76-7c6e2692aa49/volumes" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.134765 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" event={"ID":"37fe428b-b0e2-4b27-b91d-e8407e99bf81","Type":"ContainerDied","Data":"37cea62349f673df2dfd28c44750fae50d710e059455293bab1b4ddf3206bf54"} Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.134814 4741 scope.go:117] "RemoveContainer" containerID="42257dbd9289b2096acd1e52d51a22000043236c47e957f4dc8a6e60ec5997ad" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.135123 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-pfjg6" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.138720 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hqmd7" event={"ID":"2a8d0551-6f96-4651-908d-3a048f691ddb","Type":"ContainerDied","Data":"06d99d9b9a9ac9469190e4beab5a58033345d06c9af75868e9b43362404937c5"} Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.138744 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06d99d9b9a9ac9469190e4beab5a58033345d06c9af75868e9b43362404937c5" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.139100 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hqmd7" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.149945 4741 generic.go:334] "Generic (PLEG): container finished" podID="07649762-f417-4bfe-877a-1526d40f2498" containerID="8dcb2e8e8e114666df2ee7d5a78448ed7c3968bf7de7f2c22e93a9682facb037" exitCode=0 Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.150048 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sxlh8" event={"ID":"07649762-f417-4bfe-877a-1526d40f2498","Type":"ContainerDied","Data":"8dcb2e8e8e114666df2ee7d5a78448ed7c3968bf7de7f2c22e93a9682facb037"} Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.150080 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sxlh8" event={"ID":"07649762-f417-4bfe-877a-1526d40f2498","Type":"ContainerStarted","Data":"8d5b08d3c1d11d655c028172403fe99c4d8d41b8284f8e04d6146ea25ef4cec1"} Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.209848 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-pfjg6"] Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.215724 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-pfjg6"] Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.225939 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hbdwb"] Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.227442 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.230343 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hbdwb"] Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.231297 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.231447 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.234643 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.332847 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ba2de5ec-87f1-4387-befa-d853e4f877dc-etc-swift\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.333104 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-ring-data-devices\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.333175 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-combined-ca-bundle\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.333207 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdfjj\" (UniqueName: \"kubernetes.io/projected/ba2de5ec-87f1-4387-befa-d853e4f877dc-kube-api-access-kdfjj\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.333229 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-swiftconf\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.333245 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-dispersionconf\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.333302 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-scripts\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436720 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436762 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-scripts\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436800 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ba2de5ec-87f1-4387-befa-d853e4f877dc-etc-swift\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436814 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-ring-data-devices\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436867 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-combined-ca-bundle\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436896 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdfjj\" (UniqueName: \"kubernetes.io/projected/ba2de5ec-87f1-4387-befa-d853e4f877dc-kube-api-access-kdfjj\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436918 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-swiftconf\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.436933 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-dispersionconf\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.438095 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-ring-data-devices\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.438122 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-scripts\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: E0929 19:26:31.438226 4741 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 19:26:31 crc kubenswrapper[4741]: E0929 19:26:31.438239 4741 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 19:26:31 crc kubenswrapper[4741]: E0929 19:26:31.438280 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift podName:95b48711-844f-4e95-9f01-36bec6b9995a nodeName:}" failed. No retries permitted until 2025-09-29 19:26:32.438265551 +0000 UTC m=+1034.086054893 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift") pod "swift-storage-0" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a") : configmap "swift-ring-files" not found Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.438321 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ba2de5ec-87f1-4387-befa-d853e4f877dc-etc-swift\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.446860 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-swiftconf\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.448789 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-dispersionconf\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.459431 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-combined-ca-bundle\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.483232 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdfjj\" (UniqueName: \"kubernetes.io/projected/ba2de5ec-87f1-4387-befa-d853e4f877dc-kube-api-access-kdfjj\") pod \"swift-ring-rebalance-hbdwb\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.561637 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.586492 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dn5tb" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.640486 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqvpl" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.640553 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfjq5\" (UniqueName: \"kubernetes.io/projected/4eae36aa-f785-4402-ae16-3d3a68050ebb-kube-api-access-qfjq5\") pod \"4eae36aa-f785-4402-ae16-3d3a68050ebb\" (UID: \"4eae36aa-f785-4402-ae16-3d3a68050ebb\") " Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.643309 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eae36aa-f785-4402-ae16-3d3a68050ebb-kube-api-access-qfjq5" (OuterVolumeSpecName: "kube-api-access-qfjq5") pod "4eae36aa-f785-4402-ae16-3d3a68050ebb" (UID: "4eae36aa-f785-4402-ae16-3d3a68050ebb"). InnerVolumeSpecName "kube-api-access-qfjq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.742101 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pf2g\" (UniqueName: \"kubernetes.io/projected/029559ea-d2e3-4a1b-87b8-60b0a65c9730-kube-api-access-8pf2g\") pod \"029559ea-d2e3-4a1b-87b8-60b0a65c9730\" (UID: \"029559ea-d2e3-4a1b-87b8-60b0a65c9730\") " Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.742660 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfjq5\" (UniqueName: \"kubernetes.io/projected/4eae36aa-f785-4402-ae16-3d3a68050ebb-kube-api-access-qfjq5\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.747274 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/029559ea-d2e3-4a1b-87b8-60b0a65c9730-kube-api-access-8pf2g" (OuterVolumeSpecName: "kube-api-access-8pf2g") pod "029559ea-d2e3-4a1b-87b8-60b0a65c9730" (UID: "029559ea-d2e3-4a1b-87b8-60b0a65c9730"). InnerVolumeSpecName "kube-api-access-8pf2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:31 crc kubenswrapper[4741]: I0929 19:26:31.843472 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pf2g\" (UniqueName: \"kubernetes.io/projected/029559ea-d2e3-4a1b-87b8-60b0a65c9730-kube-api-access-8pf2g\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.040002 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hbdwb"] Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.157675 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dn5tb" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.157670 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dn5tb" event={"ID":"4eae36aa-f785-4402-ae16-3d3a68050ebb","Type":"ContainerDied","Data":"bf5e8a6d3d4a1f472ba723f77ec00d996861c767e8edf252351f90a06ab2c23c"} Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.158077 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf5e8a6d3d4a1f472ba723f77ec00d996861c767e8edf252351f90a06ab2c23c" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.166228 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" event={"ID":"a5a18fae-a329-44dd-9bf9-75944ef94b62","Type":"ContainerStarted","Data":"e1814248f2cfcadd9ec9b80c37fd123b0e524e173a67dd1f194d2dcadf8c47cc"} Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.166432 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.168993 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sxlh8" event={"ID":"07649762-f417-4bfe-877a-1526d40f2498","Type":"ContainerStarted","Data":"5da6e4afbcf518eac9d8729284ae631a8a8d08c43a7f4a87fd06e085bdd83378"} Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.169129 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.171179 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fqvpl" event={"ID":"029559ea-d2e3-4a1b-87b8-60b0a65c9730","Type":"ContainerDied","Data":"912dd288865183d61a5c6614c7a6a6a915eb8733a27f506cb408d03a5ff1d3a8"} Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.171221 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="912dd288865183d61a5c6614c7a6a6a915eb8733a27f506cb408d03a5ff1d3a8" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.171282 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqvpl" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.173904 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hbdwb" event={"ID":"ba2de5ec-87f1-4387-befa-d853e4f877dc","Type":"ContainerStarted","Data":"31639fdbf94bfb7b03c81b50be3f9d719a0a3f38c99369c407ff8803ea3d594b"} Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.216611 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" podStartSLOduration=4.216596269 podStartE2EDuration="4.216596269s" podCreationTimestamp="2025-09-29 19:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:26:32.194128026 +0000 UTC m=+1033.841917368" watchObservedRunningTime="2025-09-29 19:26:32.216596269 +0000 UTC m=+1033.864385601" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.216924 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-sxlh8" podStartSLOduration=3.216920189 podStartE2EDuration="3.216920189s" podCreationTimestamp="2025-09-29 19:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:26:32.215250806 +0000 UTC m=+1033.863040138" watchObservedRunningTime="2025-09-29 19:26:32.216920189 +0000 UTC m=+1033.864709521" Sep 29 19:26:32 crc kubenswrapper[4741]: I0929 19:26:32.451104 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:32 crc kubenswrapper[4741]: E0929 19:26:32.451335 4741 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 19:26:32 crc kubenswrapper[4741]: E0929 19:26:32.451381 4741 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 19:26:32 crc kubenswrapper[4741]: E0929 19:26:32.451467 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift podName:95b48711-844f-4e95-9f01-36bec6b9995a nodeName:}" failed. No retries permitted until 2025-09-29 19:26:34.451444041 +0000 UTC m=+1036.099233373 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift") pod "swift-storage-0" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a") : configmap "swift-ring-files" not found Sep 29 19:26:33 crc kubenswrapper[4741]: I0929 19:26:33.095809 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37fe428b-b0e2-4b27-b91d-e8407e99bf81" path="/var/lib/kubelet/pods/37fe428b-b0e2-4b27-b91d-e8407e99bf81/volumes" Sep 29 19:26:33 crc kubenswrapper[4741]: I0929 19:26:33.193159 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt" event={"ID":"865e963c-87bd-45aa-a4a7-95ae24dd0058","Type":"ContainerStarted","Data":"1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0"} Sep 29 19:26:33 crc kubenswrapper[4741]: I0929 19:26:33.887029 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-68snt" Sep 29 19:26:34 crc kubenswrapper[4741]: I0929 19:26:34.482238 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:34 crc kubenswrapper[4741]: E0929 19:26:34.482492 4741 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 19:26:34 crc kubenswrapper[4741]: E0929 19:26:34.482508 4741 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 19:26:34 crc kubenswrapper[4741]: E0929 19:26:34.482554 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift podName:95b48711-844f-4e95-9f01-36bec6b9995a nodeName:}" failed. No retries permitted until 2025-09-29 19:26:38.482537503 +0000 UTC m=+1040.130326835 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift") pod "swift-storage-0" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a") : configmap "swift-ring-files" not found Sep 29 19:26:35 crc kubenswrapper[4741]: I0929 19:26:35.106070 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-68snt" podStartSLOduration=14.902551704 podStartE2EDuration="42.106045788s" podCreationTimestamp="2025-09-29 19:25:53 +0000 UTC" firstStartedPulling="2025-09-29 19:26:05.335581177 +0000 UTC m=+1006.983370519" lastFinishedPulling="2025-09-29 19:26:32.539075271 +0000 UTC m=+1034.186864603" observedRunningTime="2025-09-29 19:26:33.216945582 +0000 UTC m=+1034.864734934" watchObservedRunningTime="2025-09-29 19:26:35.106045788 +0000 UTC m=+1036.753835130" Sep 29 19:26:36 crc kubenswrapper[4741]: I0929 19:26:36.218832 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84df25f5-0db8-457f-b964-3cdb332ebf02","Type":"ContainerStarted","Data":"69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000"} Sep 29 19:26:36 crc kubenswrapper[4741]: I0929 19:26:36.219361 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 29 19:26:36 crc kubenswrapper[4741]: I0929 19:26:36.220030 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hbdwb" event={"ID":"ba2de5ec-87f1-4387-befa-d853e4f877dc","Type":"ContainerStarted","Data":"80996d6bbd486a08fd7d5acf00412689ada3eb81d2f0bd189a0667e966accbcc"} Sep 29 19:26:36 crc kubenswrapper[4741]: I0929 19:26:36.242468 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.914952443 podStartE2EDuration="47.24245176s" podCreationTimestamp="2025-09-29 19:25:49 +0000 UTC" firstStartedPulling="2025-09-29 19:26:05.191963681 +0000 UTC m=+1006.839753013" lastFinishedPulling="2025-09-29 19:26:35.519462998 +0000 UTC m=+1037.167252330" observedRunningTime="2025-09-29 19:26:36.236890853 +0000 UTC m=+1037.884680185" watchObservedRunningTime="2025-09-29 19:26:36.24245176 +0000 UTC m=+1037.890241082" Sep 29 19:26:36 crc kubenswrapper[4741]: I0929 19:26:36.252627 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-hbdwb" podStartSLOduration=1.775923768 podStartE2EDuration="5.252609003s" podCreationTimestamp="2025-09-29 19:26:31 +0000 UTC" firstStartedPulling="2025-09-29 19:26:32.049594019 +0000 UTC m=+1033.697383351" lastFinishedPulling="2025-09-29 19:26:35.526279254 +0000 UTC m=+1037.174068586" observedRunningTime="2025-09-29 19:26:36.251312571 +0000 UTC m=+1037.899101913" watchObservedRunningTime="2025-09-29 19:26:36.252609003 +0000 UTC m=+1037.900398335" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.521254 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-81dc-account-create-ghmk4"] Sep 29 19:26:37 crc kubenswrapper[4741]: E0929 19:26:37.523103 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eae36aa-f785-4402-ae16-3d3a68050ebb" containerName="mariadb-database-create" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.523256 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eae36aa-f785-4402-ae16-3d3a68050ebb" containerName="mariadb-database-create" Sep 29 19:26:37 crc kubenswrapper[4741]: E0929 19:26:37.523381 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029559ea-d2e3-4a1b-87b8-60b0a65c9730" containerName="mariadb-database-create" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.523512 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="029559ea-d2e3-4a1b-87b8-60b0a65c9730" containerName="mariadb-database-create" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.523904 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="029559ea-d2e3-4a1b-87b8-60b0a65c9730" containerName="mariadb-database-create" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.524061 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eae36aa-f785-4402-ae16-3d3a68050ebb" containerName="mariadb-database-create" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.524977 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81dc-account-create-ghmk4" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.527119 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.542213 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-81dc-account-create-ghmk4"] Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.656188 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgvlc\" (UniqueName: \"kubernetes.io/projected/d1cf524c-6534-40d5-83eb-5a0b75684622-kube-api-access-kgvlc\") pod \"keystone-81dc-account-create-ghmk4\" (UID: \"d1cf524c-6534-40d5-83eb-5a0b75684622\") " pod="openstack/keystone-81dc-account-create-ghmk4" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.758417 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgvlc\" (UniqueName: \"kubernetes.io/projected/d1cf524c-6534-40d5-83eb-5a0b75684622-kube-api-access-kgvlc\") pod \"keystone-81dc-account-create-ghmk4\" (UID: \"d1cf524c-6534-40d5-83eb-5a0b75684622\") " pod="openstack/keystone-81dc-account-create-ghmk4" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.789868 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgvlc\" (UniqueName: \"kubernetes.io/projected/d1cf524c-6534-40d5-83eb-5a0b75684622-kube-api-access-kgvlc\") pod \"keystone-81dc-account-create-ghmk4\" (UID: \"d1cf524c-6534-40d5-83eb-5a0b75684622\") " pod="openstack/keystone-81dc-account-create-ghmk4" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.842804 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81dc-account-create-ghmk4" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.951464 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1b19-account-create-clkpq"] Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.952642 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1b19-account-create-clkpq" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.957677 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Sep 29 19:26:37 crc kubenswrapper[4741]: I0929 19:26:37.968996 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1b19-account-create-clkpq"] Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.063681 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s4zm\" (UniqueName: \"kubernetes.io/projected/ccccac78-84c0-4d3b-a0b5-6d29323558f3-kube-api-access-2s4zm\") pod \"placement-1b19-account-create-clkpq\" (UID: \"ccccac78-84c0-4d3b-a0b5-6d29323558f3\") " pod="openstack/placement-1b19-account-create-clkpq" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.149064 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6af0-account-create-6pmp5"] Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.150038 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6af0-account-create-6pmp5" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.154830 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.160331 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6af0-account-create-6pmp5"] Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.164803 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s4zm\" (UniqueName: \"kubernetes.io/projected/ccccac78-84c0-4d3b-a0b5-6d29323558f3-kube-api-access-2s4zm\") pod \"placement-1b19-account-create-clkpq\" (UID: \"ccccac78-84c0-4d3b-a0b5-6d29323558f3\") " pod="openstack/placement-1b19-account-create-clkpq" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.201446 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s4zm\" (UniqueName: \"kubernetes.io/projected/ccccac78-84c0-4d3b-a0b5-6d29323558f3-kube-api-access-2s4zm\") pod \"placement-1b19-account-create-clkpq\" (UID: \"ccccac78-84c0-4d3b-a0b5-6d29323558f3\") " pod="openstack/placement-1b19-account-create-clkpq" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.280348 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldvr\" (UniqueName: \"kubernetes.io/projected/f830c8bb-5f61-4a1f-bb46-d59a64d110f9-kube-api-access-bldvr\") pod \"glance-6af0-account-create-6pmp5\" (UID: \"f830c8bb-5f61-4a1f-bb46-d59a64d110f9\") " pod="openstack/glance-6af0-account-create-6pmp5" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.280803 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1b19-account-create-clkpq" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.318599 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-81dc-account-create-ghmk4"] Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.382766 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldvr\" (UniqueName: \"kubernetes.io/projected/f830c8bb-5f61-4a1f-bb46-d59a64d110f9-kube-api-access-bldvr\") pod \"glance-6af0-account-create-6pmp5\" (UID: \"f830c8bb-5f61-4a1f-bb46-d59a64d110f9\") " pod="openstack/glance-6af0-account-create-6pmp5" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.420185 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldvr\" (UniqueName: \"kubernetes.io/projected/f830c8bb-5f61-4a1f-bb46-d59a64d110f9-kube-api-access-bldvr\") pod \"glance-6af0-account-create-6pmp5\" (UID: \"f830c8bb-5f61-4a1f-bb46-d59a64d110f9\") " pod="openstack/glance-6af0-account-create-6pmp5" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.471988 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6af0-account-create-6pmp5" Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.484335 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:38 crc kubenswrapper[4741]: E0929 19:26:38.484584 4741 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 19:26:38 crc kubenswrapper[4741]: E0929 19:26:38.484600 4741 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 19:26:38 crc kubenswrapper[4741]: E0929 19:26:38.484641 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift podName:95b48711-844f-4e95-9f01-36bec6b9995a nodeName:}" failed. No retries permitted until 2025-09-29 19:26:46.484627803 +0000 UTC m=+1048.132417135 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift") pod "swift-storage-0" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a") : configmap "swift-ring-files" not found Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.803777 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1b19-account-create-clkpq"] Sep 29 19:26:38 crc kubenswrapper[4741]: W0929 19:26:38.819800 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccccac78_84c0_4d3b_a0b5_6d29323558f3.slice/crio-9eb1060987fa6d83b15267d2105a74e764d3c01b6ebb68bd7c19caf95adc8ccb WatchSource:0}: Error finding container 9eb1060987fa6d83b15267d2105a74e764d3c01b6ebb68bd7c19caf95adc8ccb: Status 404 returned error can't find the container with id 9eb1060987fa6d83b15267d2105a74e764d3c01b6ebb68bd7c19caf95adc8ccb Sep 29 19:26:38 crc kubenswrapper[4741]: I0929 19:26:38.974640 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6af0-account-create-6pmp5"] Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.171371 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.248506 4741 generic.go:334] "Generic (PLEG): container finished" podID="d1cf524c-6534-40d5-83eb-5a0b75684622" containerID="5cecc8c2ec11165c95aa8393dacbd5b4792e48d4a72193d8c53722290b713ee7" exitCode=0 Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.248620 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81dc-account-create-ghmk4" event={"ID":"d1cf524c-6534-40d5-83eb-5a0b75684622","Type":"ContainerDied","Data":"5cecc8c2ec11165c95aa8393dacbd5b4792e48d4a72193d8c53722290b713ee7"} Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.248837 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81dc-account-create-ghmk4" event={"ID":"d1cf524c-6534-40d5-83eb-5a0b75684622","Type":"ContainerStarted","Data":"94d05c9390042b1300293a040d6b7668c6764048acee0352c81f79f6f5c8dc35"} Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.250675 4741 generic.go:334] "Generic (PLEG): container finished" podID="ccccac78-84c0-4d3b-a0b5-6d29323558f3" containerID="915657e3007151e1d5101dbb2d89c3041033781338926c491ec154e4b25f1c9a" exitCode=0 Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.250740 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1b19-account-create-clkpq" event={"ID":"ccccac78-84c0-4d3b-a0b5-6d29323558f3","Type":"ContainerDied","Data":"915657e3007151e1d5101dbb2d89c3041033781338926c491ec154e4b25f1c9a"} Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.250768 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1b19-account-create-clkpq" event={"ID":"ccccac78-84c0-4d3b-a0b5-6d29323558f3","Type":"ContainerStarted","Data":"9eb1060987fa6d83b15267d2105a74e764d3c01b6ebb68bd7c19caf95adc8ccb"} Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.252531 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6af0-account-create-6pmp5" event={"ID":"f830c8bb-5f61-4a1f-bb46-d59a64d110f9","Type":"ContainerStarted","Data":"d21304531a9d12f50d20982ea29f8c74f91db31217925de0d1242446471f527d"} Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.252565 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6af0-account-create-6pmp5" event={"ID":"f830c8bb-5f61-4a1f-bb46-d59a64d110f9","Type":"ContainerStarted","Data":"c64401a1450fcf939d2c8062e6fe759d627af2f9acbbf0ef00376df172074d2e"} Sep 29 19:26:39 crc kubenswrapper[4741]: I0929 19:26:39.282912 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-6af0-account-create-6pmp5" podStartSLOduration=1.282897317 podStartE2EDuration="1.282897317s" podCreationTimestamp="2025-09-29 19:26:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:26:39.279371506 +0000 UTC m=+1040.927160838" watchObservedRunningTime="2025-09-29 19:26:39.282897317 +0000 UTC m=+1040.930686649" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.005532 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.063607 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tvwnx"] Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.063828 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerName="dnsmasq-dns" containerID="cri-o://e1814248f2cfcadd9ec9b80c37fd123b0e524e173a67dd1f194d2dcadf8c47cc" gracePeriod=10 Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.259874 4741 generic.go:334] "Generic (PLEG): container finished" podID="f830c8bb-5f61-4a1f-bb46-d59a64d110f9" containerID="d21304531a9d12f50d20982ea29f8c74f91db31217925de0d1242446471f527d" exitCode=0 Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.260207 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6af0-account-create-6pmp5" event={"ID":"f830c8bb-5f61-4a1f-bb46-d59a64d110f9","Type":"ContainerDied","Data":"d21304531a9d12f50d20982ea29f8c74f91db31217925de0d1242446471f527d"} Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.271289 4741 generic.go:334] "Generic (PLEG): container finished" podID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerID="e1814248f2cfcadd9ec9b80c37fd123b0e524e173a67dd1f194d2dcadf8c47cc" exitCode=0 Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.271376 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" event={"ID":"a5a18fae-a329-44dd-9bf9-75944ef94b62","Type":"ContainerDied","Data":"e1814248f2cfcadd9ec9b80c37fd123b0e524e173a67dd1f194d2dcadf8c47cc"} Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.502775 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.613469 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81dc-account-create-ghmk4" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.627744 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1b19-account-create-clkpq" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.634027 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-config\") pod \"a5a18fae-a329-44dd-9bf9-75944ef94b62\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.634167 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-dns-svc\") pod \"a5a18fae-a329-44dd-9bf9-75944ef94b62\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.634209 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvpxc\" (UniqueName: \"kubernetes.io/projected/a5a18fae-a329-44dd-9bf9-75944ef94b62-kube-api-access-kvpxc\") pod \"a5a18fae-a329-44dd-9bf9-75944ef94b62\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.634225 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-nb\") pod \"a5a18fae-a329-44dd-9bf9-75944ef94b62\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.634345 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-sb\") pod \"a5a18fae-a329-44dd-9bf9-75944ef94b62\" (UID: \"a5a18fae-a329-44dd-9bf9-75944ef94b62\") " Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.639701 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a18fae-a329-44dd-9bf9-75944ef94b62-kube-api-access-kvpxc" (OuterVolumeSpecName: "kube-api-access-kvpxc") pod "a5a18fae-a329-44dd-9bf9-75944ef94b62" (UID: "a5a18fae-a329-44dd-9bf9-75944ef94b62"). InnerVolumeSpecName "kube-api-access-kvpxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.682435 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-config" (OuterVolumeSpecName: "config") pod "a5a18fae-a329-44dd-9bf9-75944ef94b62" (UID: "a5a18fae-a329-44dd-9bf9-75944ef94b62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.684259 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a5a18fae-a329-44dd-9bf9-75944ef94b62" (UID: "a5a18fae-a329-44dd-9bf9-75944ef94b62"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.687867 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a5a18fae-a329-44dd-9bf9-75944ef94b62" (UID: "a5a18fae-a329-44dd-9bf9-75944ef94b62"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.690294 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a5a18fae-a329-44dd-9bf9-75944ef94b62" (UID: "a5a18fae-a329-44dd-9bf9-75944ef94b62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.736552 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgvlc\" (UniqueName: \"kubernetes.io/projected/d1cf524c-6534-40d5-83eb-5a0b75684622-kube-api-access-kgvlc\") pod \"d1cf524c-6534-40d5-83eb-5a0b75684622\" (UID: \"d1cf524c-6534-40d5-83eb-5a0b75684622\") " Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.736871 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s4zm\" (UniqueName: \"kubernetes.io/projected/ccccac78-84c0-4d3b-a0b5-6d29323558f3-kube-api-access-2s4zm\") pod \"ccccac78-84c0-4d3b-a0b5-6d29323558f3\" (UID: \"ccccac78-84c0-4d3b-a0b5-6d29323558f3\") " Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.737275 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.737346 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.737428 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.737502 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvpxc\" (UniqueName: \"kubernetes.io/projected/a5a18fae-a329-44dd-9bf9-75944ef94b62-kube-api-access-kvpxc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.737593 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5a18fae-a329-44dd-9bf9-75944ef94b62-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.740596 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccccac78-84c0-4d3b-a0b5-6d29323558f3-kube-api-access-2s4zm" (OuterVolumeSpecName: "kube-api-access-2s4zm") pod "ccccac78-84c0-4d3b-a0b5-6d29323558f3" (UID: "ccccac78-84c0-4d3b-a0b5-6d29323558f3"). InnerVolumeSpecName "kube-api-access-2s4zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.747333 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1cf524c-6534-40d5-83eb-5a0b75684622-kube-api-access-kgvlc" (OuterVolumeSpecName: "kube-api-access-kgvlc") pod "d1cf524c-6534-40d5-83eb-5a0b75684622" (UID: "d1cf524c-6534-40d5-83eb-5a0b75684622"). InnerVolumeSpecName "kube-api-access-kgvlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.838962 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgvlc\" (UniqueName: \"kubernetes.io/projected/d1cf524c-6534-40d5-83eb-5a0b75684622-kube-api-access-kgvlc\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:40 crc kubenswrapper[4741]: I0929 19:26:40.838992 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s4zm\" (UniqueName: \"kubernetes.io/projected/ccccac78-84c0-4d3b-a0b5-6d29323558f3-kube-api-access-2s4zm\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.279526 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" event={"ID":"a5a18fae-a329-44dd-9bf9-75944ef94b62","Type":"ContainerDied","Data":"8e5fa4dcdef906ef42794f8e12034e9fc16b6bae9058e55a04cdd14f235b472b"} Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.279580 4741 scope.go:117] "RemoveContainer" containerID="e1814248f2cfcadd9ec9b80c37fd123b0e524e173a67dd1f194d2dcadf8c47cc" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.279594 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tvwnx" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.282218 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f5597514-2fd8-4d92-b115-05b66894ea94","Type":"ContainerStarted","Data":"c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f"} Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.284241 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81dc-account-create-ghmk4" event={"ID":"d1cf524c-6534-40d5-83eb-5a0b75684622","Type":"ContainerDied","Data":"94d05c9390042b1300293a040d6b7668c6764048acee0352c81f79f6f5c8dc35"} Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.284283 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94d05c9390042b1300293a040d6b7668c6764048acee0352c81f79f6f5c8dc35" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.284343 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81dc-account-create-ghmk4" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.286547 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1b19-account-create-clkpq" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.286607 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1b19-account-create-clkpq" event={"ID":"ccccac78-84c0-4d3b-a0b5-6d29323558f3","Type":"ContainerDied","Data":"9eb1060987fa6d83b15267d2105a74e764d3c01b6ebb68bd7c19caf95adc8ccb"} Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.286649 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9eb1060987fa6d83b15267d2105a74e764d3c01b6ebb68bd7c19caf95adc8ccb" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.308217 4741 scope.go:117] "RemoveContainer" containerID="d6b611a9627ae6bee6c41ccb8652ca3dc9faffd81accc380242d52cb411f881e" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.308465 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.235766717 podStartE2EDuration="46.308445263s" podCreationTimestamp="2025-09-29 19:25:55 +0000 UTC" firstStartedPulling="2025-09-29 19:26:05.446611971 +0000 UTC m=+1007.094401303" lastFinishedPulling="2025-09-29 19:26:40.519290527 +0000 UTC m=+1042.167079849" observedRunningTime="2025-09-29 19:26:41.305376036 +0000 UTC m=+1042.953165358" watchObservedRunningTime="2025-09-29 19:26:41.308445263 +0000 UTC m=+1042.956234605" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.329933 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tvwnx"] Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.336130 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tvwnx"] Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.586300 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6af0-account-create-6pmp5" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.654803 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bldvr\" (UniqueName: \"kubernetes.io/projected/f830c8bb-5f61-4a1f-bb46-d59a64d110f9-kube-api-access-bldvr\") pod \"f830c8bb-5f61-4a1f-bb46-d59a64d110f9\" (UID: \"f830c8bb-5f61-4a1f-bb46-d59a64d110f9\") " Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.659747 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f830c8bb-5f61-4a1f-bb46-d59a64d110f9-kube-api-access-bldvr" (OuterVolumeSpecName: "kube-api-access-bldvr") pod "f830c8bb-5f61-4a1f-bb46-d59a64d110f9" (UID: "f830c8bb-5f61-4a1f-bb46-d59a64d110f9"). InnerVolumeSpecName "kube-api-access-bldvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.741244 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.741286 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Sep 29 19:26:41 crc kubenswrapper[4741]: I0929 19:26:41.757162 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bldvr\" (UniqueName: \"kubernetes.io/projected/f830c8bb-5f61-4a1f-bb46-d59a64d110f9-kube-api-access-bldvr\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:42 crc kubenswrapper[4741]: I0929 19:26:42.296085 4741 generic.go:334] "Generic (PLEG): container finished" podID="ba2de5ec-87f1-4387-befa-d853e4f877dc" containerID="80996d6bbd486a08fd7d5acf00412689ada3eb81d2f0bd189a0667e966accbcc" exitCode=0 Sep 29 19:26:42 crc kubenswrapper[4741]: I0929 19:26:42.296221 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hbdwb" event={"ID":"ba2de5ec-87f1-4387-befa-d853e4f877dc","Type":"ContainerDied","Data":"80996d6bbd486a08fd7d5acf00412689ada3eb81d2f0bd189a0667e966accbcc"} Sep 29 19:26:42 crc kubenswrapper[4741]: I0929 19:26:42.299322 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6af0-account-create-6pmp5" event={"ID":"f830c8bb-5f61-4a1f-bb46-d59a64d110f9","Type":"ContainerDied","Data":"c64401a1450fcf939d2c8062e6fe759d627af2f9acbbf0ef00376df172074d2e"} Sep 29 19:26:42 crc kubenswrapper[4741]: I0929 19:26:42.299439 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c64401a1450fcf939d2c8062e6fe759d627af2f9acbbf0ef00376df172074d2e" Sep 29 19:26:42 crc kubenswrapper[4741]: I0929 19:26:42.299563 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6af0-account-create-6pmp5" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.094279 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" path="/var/lib/kubelet/pods/a5a18fae-a329-44dd-9bf9-75944ef94b62/volumes" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295197 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-z95rt"] Sep 29 19:26:43 crc kubenswrapper[4741]: E0929 19:26:43.295600 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerName="init" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295624 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerName="init" Sep 29 19:26:43 crc kubenswrapper[4741]: E0929 19:26:43.295643 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cf524c-6534-40d5-83eb-5a0b75684622" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295653 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cf524c-6534-40d5-83eb-5a0b75684622" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: E0929 19:26:43.295676 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerName="dnsmasq-dns" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295685 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerName="dnsmasq-dns" Sep 29 19:26:43 crc kubenswrapper[4741]: E0929 19:26:43.295700 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccccac78-84c0-4d3b-a0b5-6d29323558f3" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295708 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccccac78-84c0-4d3b-a0b5-6d29323558f3" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: E0929 19:26:43.295725 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f830c8bb-5f61-4a1f-bb46-d59a64d110f9" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295732 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f830c8bb-5f61-4a1f-bb46-d59a64d110f9" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295943 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccccac78-84c0-4d3b-a0b5-6d29323558f3" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295958 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f830c8bb-5f61-4a1f-bb46-d59a64d110f9" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295978 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1cf524c-6534-40d5-83eb-5a0b75684622" containerName="mariadb-account-create" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.295998 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a18fae-a329-44dd-9bf9-75944ef94b62" containerName="dnsmasq-dns" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.296701 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.300768 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.301544 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-h9mn9" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.302751 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-z95rt"] Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.383647 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-combined-ca-bundle\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.384092 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-config-data\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.384141 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-db-sync-config-data\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.384187 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc9b6\" (UniqueName: \"kubernetes.io/projected/023a246b-31d3-47bb-822b-c8e61aa2f034-kube-api-access-vc9b6\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.485761 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc9b6\" (UniqueName: \"kubernetes.io/projected/023a246b-31d3-47bb-822b-c8e61aa2f034-kube-api-access-vc9b6\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.485852 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-combined-ca-bundle\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.485905 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-config-data\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.485936 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-db-sync-config-data\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.491952 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-config-data\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.492882 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-db-sync-config-data\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.493427 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-combined-ca-bundle\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.504822 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc9b6\" (UniqueName: \"kubernetes.io/projected/023a246b-31d3-47bb-822b-c8e61aa2f034-kube-api-access-vc9b6\") pod \"glance-db-sync-z95rt\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.634429 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z95rt" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.710071 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.790134 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdfjj\" (UniqueName: \"kubernetes.io/projected/ba2de5ec-87f1-4387-befa-d853e4f877dc-kube-api-access-kdfjj\") pod \"ba2de5ec-87f1-4387-befa-d853e4f877dc\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.790576 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-ring-data-devices\") pod \"ba2de5ec-87f1-4387-befa-d853e4f877dc\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.790647 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-scripts\") pod \"ba2de5ec-87f1-4387-befa-d853e4f877dc\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.790716 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-combined-ca-bundle\") pod \"ba2de5ec-87f1-4387-befa-d853e4f877dc\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.790782 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ba2de5ec-87f1-4387-befa-d853e4f877dc-etc-swift\") pod \"ba2de5ec-87f1-4387-befa-d853e4f877dc\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.790845 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-dispersionconf\") pod \"ba2de5ec-87f1-4387-befa-d853e4f877dc\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.790916 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-swiftconf\") pod \"ba2de5ec-87f1-4387-befa-d853e4f877dc\" (UID: \"ba2de5ec-87f1-4387-befa-d853e4f877dc\") " Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.792608 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ba2de5ec-87f1-4387-befa-d853e4f877dc" (UID: "ba2de5ec-87f1-4387-befa-d853e4f877dc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.792826 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba2de5ec-87f1-4387-befa-d853e4f877dc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ba2de5ec-87f1-4387-befa-d853e4f877dc" (UID: "ba2de5ec-87f1-4387-befa-d853e4f877dc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.795979 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba2de5ec-87f1-4387-befa-d853e4f877dc-kube-api-access-kdfjj" (OuterVolumeSpecName: "kube-api-access-kdfjj") pod "ba2de5ec-87f1-4387-befa-d853e4f877dc" (UID: "ba2de5ec-87f1-4387-befa-d853e4f877dc"). InnerVolumeSpecName "kube-api-access-kdfjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.802740 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ba2de5ec-87f1-4387-befa-d853e4f877dc" (UID: "ba2de5ec-87f1-4387-befa-d853e4f877dc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.819635 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba2de5ec-87f1-4387-befa-d853e4f877dc" (UID: "ba2de5ec-87f1-4387-befa-d853e4f877dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.823143 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-scripts" (OuterVolumeSpecName: "scripts") pod "ba2de5ec-87f1-4387-befa-d853e4f877dc" (UID: "ba2de5ec-87f1-4387-befa-d853e4f877dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.826713 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ba2de5ec-87f1-4387-befa-d853e4f877dc" (UID: "ba2de5ec-87f1-4387-befa-d853e4f877dc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.892612 4741 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.892669 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdfjj\" (UniqueName: \"kubernetes.io/projected/ba2de5ec-87f1-4387-befa-d853e4f877dc-kube-api-access-kdfjj\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.892683 4741 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.892692 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba2de5ec-87f1-4387-befa-d853e4f877dc-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.892700 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.892708 4741 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ba2de5ec-87f1-4387-befa-d853e4f877dc-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:43 crc kubenswrapper[4741]: I0929 19:26:43.892734 4741 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ba2de5ec-87f1-4387-befa-d853e4f877dc-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 29 19:26:44 crc kubenswrapper[4741]: I0929 19:26:44.156879 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-z95rt"] Sep 29 19:26:44 crc kubenswrapper[4741]: I0929 19:26:44.316723 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z95rt" event={"ID":"023a246b-31d3-47bb-822b-c8e61aa2f034","Type":"ContainerStarted","Data":"9f436e855819643041a51daed5ec88be1ac9c54b7d3d6d76bff7d21c2619e008"} Sep 29 19:26:44 crc kubenswrapper[4741]: I0929 19:26:44.318327 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hbdwb" event={"ID":"ba2de5ec-87f1-4387-befa-d853e4f877dc","Type":"ContainerDied","Data":"31639fdbf94bfb7b03c81b50be3f9d719a0a3f38c99369c407ff8803ea3d594b"} Sep 29 19:26:44 crc kubenswrapper[4741]: I0929 19:26:44.318358 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31639fdbf94bfb7b03c81b50be3f9d719a0a3f38c99369c407ff8803ea3d594b" Sep 29 19:26:44 crc kubenswrapper[4741]: I0929 19:26:44.318471 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hbdwb" Sep 29 19:26:44 crc kubenswrapper[4741]: I0929 19:26:44.784552 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Sep 29 19:26:46 crc kubenswrapper[4741]: I0929 19:26:46.532927 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:46 crc kubenswrapper[4741]: I0929 19:26:46.552839 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"swift-storage-0\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " pod="openstack/swift-storage-0" Sep 29 19:26:46 crc kubenswrapper[4741]: I0929 19:26:46.780736 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Sep 29 19:26:46 crc kubenswrapper[4741]: I0929 19:26:46.830840 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.152283 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Sep 29 19:26:47 crc kubenswrapper[4741]: E0929 19:26:47.153020 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2de5ec-87f1-4387-befa-d853e4f877dc" containerName="swift-ring-rebalance" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.153037 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2de5ec-87f1-4387-befa-d853e4f877dc" containerName="swift-ring-rebalance" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.153229 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba2de5ec-87f1-4387-befa-d853e4f877dc" containerName="swift-ring-rebalance" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.155281 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.159664 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.159816 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.161434 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.161668 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-mhmx7" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.163041 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.268362 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.268599 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cj96\" (UniqueName: \"kubernetes.io/projected/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-kube-api-access-2cj96\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.268631 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-config\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.268823 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.268887 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-scripts\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.268914 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.268959 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.370065 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.370427 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-scripts\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.370469 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.370486 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.370542 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.370572 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cj96\" (UniqueName: \"kubernetes.io/projected/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-kube-api-access-2cj96\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.370599 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-config\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.371590 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.371972 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-scripts\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.372070 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-config\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.378031 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.378748 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.384279 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.386876 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cj96\" (UniqueName: \"kubernetes.io/projected/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-kube-api-access-2cj96\") pod \"ovn-northd-0\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.499261 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.551523 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 29 19:26:47 crc kubenswrapper[4741]: W0929 19:26:47.560150 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95b48711_844f_4e95_9f01_36bec6b9995a.slice/crio-d1de83713e8780fdcbf82c2f2fcedf1895acdd57776e89f880ce3263918c9aac WatchSource:0}: Error finding container d1de83713e8780fdcbf82c2f2fcedf1895acdd57776e89f880ce3263918c9aac: Status 404 returned error can't find the container with id d1de83713e8780fdcbf82c2f2fcedf1895acdd57776e89f880ce3263918c9aac Sep 29 19:26:47 crc kubenswrapper[4741]: I0929 19:26:47.936926 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 19:26:47 crc kubenswrapper[4741]: W0929 19:26:47.941558 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d4bbc33_c647_4eda_bdc6_89f80678ae8d.slice/crio-7a4cadaa08ec960786e08317688380e6c5819fa220f871529e224f3445c214e1 WatchSource:0}: Error finding container 7a4cadaa08ec960786e08317688380e6c5819fa220f871529e224f3445c214e1: Status 404 returned error can't find the container with id 7a4cadaa08ec960786e08317688380e6c5819fa220f871529e224f3445c214e1 Sep 29 19:26:48 crc kubenswrapper[4741]: I0929 19:26:48.346745 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9d4bbc33-c647-4eda-bdc6-89f80678ae8d","Type":"ContainerStarted","Data":"7a4cadaa08ec960786e08317688380e6c5819fa220f871529e224f3445c214e1"} Sep 29 19:26:48 crc kubenswrapper[4741]: I0929 19:26:48.348554 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"d1de83713e8780fdcbf82c2f2fcedf1895acdd57776e89f880ce3263918c9aac"} Sep 29 19:26:49 crc kubenswrapper[4741]: I0929 19:26:49.636103 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 29 19:26:52 crc kubenswrapper[4741]: I0929 19:26:52.380196 4741 generic.go:334] "Generic (PLEG): container finished" podID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerID="7b27d2ea63000a29d7fab7a6fb5ae477f330f125488a691cbc9155958f46378e" exitCode=0 Sep 29 19:26:52 crc kubenswrapper[4741]: I0929 19:26:52.380265 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8e0c02dc-69a9-4e60-b179-0e23842d10a4","Type":"ContainerDied","Data":"7b27d2ea63000a29d7fab7a6fb5ae477f330f125488a691cbc9155958f46378e"} Sep 29 19:26:52 crc kubenswrapper[4741]: I0929 19:26:52.383139 4741 generic.go:334] "Generic (PLEG): container finished" podID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerID="298a19b9e2f8e061524da9a8a4b1326a3050a055dc372e86be58a14f3e42e0e3" exitCode=0 Sep 29 19:26:52 crc kubenswrapper[4741]: I0929 19:26:52.383164 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df6beb49-03ad-47ef-a9c7-3f37baa6d105","Type":"ContainerDied","Data":"298a19b9e2f8e061524da9a8a4b1326a3050a055dc372e86be58a14f3e42e0e3"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.433663 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9d4bbc33-c647-4eda-bdc6-89f80678ae8d","Type":"ContainerStarted","Data":"e2974921beee8fdb66bfc522f1bf2e042117a87afcc7cd9f689c79ab47e6bcf2"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.434212 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9d4bbc33-c647-4eda-bdc6-89f80678ae8d","Type":"ContainerStarted","Data":"3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.434231 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.434892 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z95rt" event={"ID":"023a246b-31d3-47bb-822b-c8e61aa2f034","Type":"ContainerStarted","Data":"59ff3431a84ba005bd2b5c3d863965362c61aebed601f9c5ea3a2b72e47dc82b"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.437377 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"5c99d27cbf271df86190fc67b6bd3129d7c9ebc403240020a83b932781c09aa8"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.437427 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"bdcc643cf7b0ba52293b398520b054d5a0af7419de83697c00319b14e82a22f5"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.437436 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"0f18bf31b15ac6485b4321250e71d201fadbe48a440243ff5584feadda173ffc"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.439839 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8e0c02dc-69a9-4e60-b179-0e23842d10a4","Type":"ContainerStarted","Data":"7d7103d33c61d8d8489afa8a3b1ea5db27133d9afa53b11454cfade2f1762a5c"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.440037 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.441179 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df6beb49-03ad-47ef-a9c7-3f37baa6d105","Type":"ContainerStarted","Data":"05c981478a21299cdffcc57230be7582dcc6e046d2f10f197e1a1fb4606f05e6"} Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.441583 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.456479 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.7580984979999998 podStartE2EDuration="8.456462374s" podCreationTimestamp="2025-09-29 19:26:47 +0000 UTC" firstStartedPulling="2025-09-29 19:26:47.944152738 +0000 UTC m=+1049.591942070" lastFinishedPulling="2025-09-29 19:26:54.642516614 +0000 UTC m=+1056.290305946" observedRunningTime="2025-09-29 19:26:55.45254039 +0000 UTC m=+1057.100329732" watchObservedRunningTime="2025-09-29 19:26:55.456462374 +0000 UTC m=+1057.104251706" Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.477502 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-z95rt" podStartSLOduration=1.999544529 podStartE2EDuration="12.477485079s" podCreationTimestamp="2025-09-29 19:26:43 +0000 UTC" firstStartedPulling="2025-09-29 19:26:44.164490441 +0000 UTC m=+1045.812279773" lastFinishedPulling="2025-09-29 19:26:54.642430991 +0000 UTC m=+1056.290220323" observedRunningTime="2025-09-29 19:26:55.471709616 +0000 UTC m=+1057.119498948" watchObservedRunningTime="2025-09-29 19:26:55.477485079 +0000 UTC m=+1057.125274411" Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.538169 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.464010442 podStartE2EDuration="1m12.538149117s" podCreationTimestamp="2025-09-29 19:25:43 +0000 UTC" firstStartedPulling="2025-09-29 19:25:45.231957914 +0000 UTC m=+986.879747246" lastFinishedPulling="2025-09-29 19:26:21.306096589 +0000 UTC m=+1022.953885921" observedRunningTime="2025-09-29 19:26:55.511268257 +0000 UTC m=+1057.159057589" watchObservedRunningTime="2025-09-29 19:26:55.538149117 +0000 UTC m=+1057.185938459" Sep 29 19:26:55 crc kubenswrapper[4741]: I0929 19:26:55.538798 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.459134978 podStartE2EDuration="1m12.538790888s" podCreationTimestamp="2025-09-29 19:25:43 +0000 UTC" firstStartedPulling="2025-09-29 19:25:45.414224478 +0000 UTC m=+987.062013820" lastFinishedPulling="2025-09-29 19:26:21.493880408 +0000 UTC m=+1023.141669730" observedRunningTime="2025-09-29 19:26:55.53478389 +0000 UTC m=+1057.182573222" watchObservedRunningTime="2025-09-29 19:26:55.538790888 +0000 UTC m=+1057.186580220" Sep 29 19:26:56 crc kubenswrapper[4741]: I0929 19:26:56.452122 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"11c745f1505ad37096a8441c7db5e1f99996efc6719e0f171773f7647d65d6f0"} Sep 29 19:26:57 crc kubenswrapper[4741]: I0929 19:26:57.462240 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"49334c9dcc9f0ab359baa344e3fe68ec6f0e5271440eec25dd046555f7cc0e20"} Sep 29 19:26:57 crc kubenswrapper[4741]: I0929 19:26:57.462685 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"8bdeba0cf6a202fe00074a937a1baa94541c1fc21049258f466a2ee65af3a75e"} Sep 29 19:26:57 crc kubenswrapper[4741]: I0929 19:26:57.462698 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"e650c5ffc6ebc6918dc63202d7b9d6aaff80d4b6db1471fd84f3faebfd8b934f"} Sep 29 19:26:57 crc kubenswrapper[4741]: I0929 19:26:57.462707 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"0272c3b7736bc9733f76397bbd76a46beab031e046ac627b3f5b05578332ba13"} Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.042669 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.055614 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.277870 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-68snt-config-6qk62"] Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.278827 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.280516 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.290808 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-68snt-config-6qk62"] Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.394052 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.394121 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run-ovn\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.394172 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-scripts\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.394204 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmv5v\" (UniqueName: \"kubernetes.io/projected/4e24c598-1b77-4133-8840-4e91c8040762-kube-api-access-fmv5v\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.394500 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-log-ovn\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.394584 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-additional-scripts\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.482509 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"878b7f5b4bcf11108cbc1b12fcff16a252c21bb2290b20babdfe2cc60f5f4278"} Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.482574 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"a68bac92910f13da6b5bd1c4c6fa29f7fec8ba0ea712f3e3c9da7813f47eafee"} Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.482593 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"5ace69d8ba764865fefb3a17c320882c81b7b6f3fba4329e49c21e37a73ff7f1"} Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.482605 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"71e20963cd9fad0ee127b4dda01c84dd767efbefb2fcb80296923f8718f1ddbf"} Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.496222 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-log-ovn\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.496291 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-additional-scripts\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.496369 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.496440 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run-ovn\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.496489 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-scripts\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.496534 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmv5v\" (UniqueName: \"kubernetes.io/projected/4e24c598-1b77-4133-8840-4e91c8040762-kube-api-access-fmv5v\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.497206 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-log-ovn\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.498146 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-additional-scripts\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.498222 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.498273 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run-ovn\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.500648 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-scripts\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.513590 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmv5v\" (UniqueName: \"kubernetes.io/projected/4e24c598-1b77-4133-8840-4e91c8040762-kube-api-access-fmv5v\") pod \"ovn-controller-68snt-config-6qk62\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:26:59 crc kubenswrapper[4741]: I0929 19:26:59.594009 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.129405 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-68snt-config-6qk62"] Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.492426 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt-config-6qk62" event={"ID":"4e24c598-1b77-4133-8840-4e91c8040762","Type":"ContainerStarted","Data":"1176f4d9203ffd34279ab2077789bd46d8d3ccd54e327e9be6c17203056b22cf"} Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.492759 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt-config-6qk62" event={"ID":"4e24c598-1b77-4133-8840-4e91c8040762","Type":"ContainerStarted","Data":"7e1269ed7318deec80f17ebb149b7bc8720dfce0b4d38f763677f498d06f11a1"} Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.500413 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"e8277ed6dfec4fbba8595886bec52f0c929d6c4a577686a175df2b460882c90f"} Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.500467 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"af59f98d013533f7b1f00c005e49813bc7d0e7ab704e34a34db4cec3151d2b45"} Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.500477 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerStarted","Data":"a979b22c48d4833060b798e59c6172ae0d0fb020aa7081564218be678d051ef8"} Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.514232 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-68snt-config-6qk62" podStartSLOduration=1.514212529 podStartE2EDuration="1.514212529s" podCreationTimestamp="2025-09-29 19:26:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:00.513243217 +0000 UTC m=+1062.161032559" watchObservedRunningTime="2025-09-29 19:27:00.514212529 +0000 UTC m=+1062.162001861" Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.555255 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.42443369 podStartE2EDuration="31.555241176s" podCreationTimestamp="2025-09-29 19:26:29 +0000 UTC" firstStartedPulling="2025-09-29 19:26:47.563733248 +0000 UTC m=+1049.211522580" lastFinishedPulling="2025-09-29 19:26:58.694540734 +0000 UTC m=+1060.342330066" observedRunningTime="2025-09-29 19:27:00.548970067 +0000 UTC m=+1062.196759389" watchObservedRunningTime="2025-09-29 19:27:00.555241176 +0000 UTC m=+1062.203030508" Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.872407 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-m9k4f"] Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.873656 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.875877 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Sep 29 19:27:00 crc kubenswrapper[4741]: I0929 19:27:00.907988 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-m9k4f"] Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.025282 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.025319 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-config\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.025338 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcfgg\" (UniqueName: \"kubernetes.io/projected/99b291b9-b83d-4854-aa81-3ca6ab88cbee-kube-api-access-lcfgg\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.025546 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.025596 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-svc\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.025975 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.127755 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-config\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.127812 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcfgg\" (UniqueName: \"kubernetes.io/projected/99b291b9-b83d-4854-aa81-3ca6ab88cbee-kube-api-access-lcfgg\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.127889 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.127913 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-svc\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.128018 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.128046 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.128655 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-config\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.128988 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.129116 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.129365 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-svc\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.129782 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.147309 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcfgg\" (UniqueName: \"kubernetes.io/projected/99b291b9-b83d-4854-aa81-3ca6ab88cbee-kube-api-access-lcfgg\") pod \"dnsmasq-dns-764c5664d7-m9k4f\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.191821 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.444544 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-m9k4f"] Sep 29 19:27:01 crc kubenswrapper[4741]: W0929 19:27:01.458244 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99b291b9_b83d_4854_aa81_3ca6ab88cbee.slice/crio-5ffc6de5063caa019acd16d2f477e0b11cb997f9a4db647eccf5069959418bd8 WatchSource:0}: Error finding container 5ffc6de5063caa019acd16d2f477e0b11cb997f9a4db647eccf5069959418bd8: Status 404 returned error can't find the container with id 5ffc6de5063caa019acd16d2f477e0b11cb997f9a4db647eccf5069959418bd8 Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.530173 4741 generic.go:334] "Generic (PLEG): container finished" podID="4e24c598-1b77-4133-8840-4e91c8040762" containerID="1176f4d9203ffd34279ab2077789bd46d8d3ccd54e327e9be6c17203056b22cf" exitCode=0 Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.530646 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt-config-6qk62" event={"ID":"4e24c598-1b77-4133-8840-4e91c8040762","Type":"ContainerDied","Data":"1176f4d9203ffd34279ab2077789bd46d8d3ccd54e327e9be6c17203056b22cf"} Sep 29 19:27:01 crc kubenswrapper[4741]: I0929 19:27:01.534124 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" event={"ID":"99b291b9-b83d-4854-aa81-3ca6ab88cbee","Type":"ContainerStarted","Data":"5ffc6de5063caa019acd16d2f477e0b11cb997f9a4db647eccf5069959418bd8"} Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.541219 4741 generic.go:334] "Generic (PLEG): container finished" podID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerID="390d29518c9894105621a83295f973d173b7f1e39264f098ed6df915ad191c2b" exitCode=0 Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.541299 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" event={"ID":"99b291b9-b83d-4854-aa81-3ca6ab88cbee","Type":"ContainerDied","Data":"390d29518c9894105621a83295f973d173b7f1e39264f098ed6df915ad191c2b"} Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.886070 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954311 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmv5v\" (UniqueName: \"kubernetes.io/projected/4e24c598-1b77-4133-8840-4e91c8040762-kube-api-access-fmv5v\") pod \"4e24c598-1b77-4133-8840-4e91c8040762\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954367 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-log-ovn\") pod \"4e24c598-1b77-4133-8840-4e91c8040762\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954435 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run-ovn\") pod \"4e24c598-1b77-4133-8840-4e91c8040762\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954465 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run\") pod \"4e24c598-1b77-4133-8840-4e91c8040762\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954511 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-scripts\") pod \"4e24c598-1b77-4133-8840-4e91c8040762\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954521 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4e24c598-1b77-4133-8840-4e91c8040762" (UID: "4e24c598-1b77-4133-8840-4e91c8040762"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954582 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4e24c598-1b77-4133-8840-4e91c8040762" (UID: "4e24c598-1b77-4133-8840-4e91c8040762"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954590 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-additional-scripts\") pod \"4e24c598-1b77-4133-8840-4e91c8040762\" (UID: \"4e24c598-1b77-4133-8840-4e91c8040762\") " Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954601 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run" (OuterVolumeSpecName: "var-run") pod "4e24c598-1b77-4133-8840-4e91c8040762" (UID: "4e24c598-1b77-4133-8840-4e91c8040762"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954889 4741 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954905 4741 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.954914 4741 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e24c598-1b77-4133-8840-4e91c8040762-var-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.955409 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4e24c598-1b77-4133-8840-4e91c8040762" (UID: "4e24c598-1b77-4133-8840-4e91c8040762"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.955552 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-scripts" (OuterVolumeSpecName: "scripts") pod "4e24c598-1b77-4133-8840-4e91c8040762" (UID: "4e24c598-1b77-4133-8840-4e91c8040762"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:02 crc kubenswrapper[4741]: I0929 19:27:02.961080 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e24c598-1b77-4133-8840-4e91c8040762-kube-api-access-fmv5v" (OuterVolumeSpecName: "kube-api-access-fmv5v") pod "4e24c598-1b77-4133-8840-4e91c8040762" (UID: "4e24c598-1b77-4133-8840-4e91c8040762"). InnerVolumeSpecName "kube-api-access-fmv5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.056287 4741 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.056329 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmv5v\" (UniqueName: \"kubernetes.io/projected/4e24c598-1b77-4133-8840-4e91c8040762-kube-api-access-fmv5v\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.056356 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e24c598-1b77-4133-8840-4e91c8040762-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.549613 4741 generic.go:334] "Generic (PLEG): container finished" podID="023a246b-31d3-47bb-822b-c8e61aa2f034" containerID="59ff3431a84ba005bd2b5c3d863965362c61aebed601f9c5ea3a2b72e47dc82b" exitCode=0 Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.549690 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z95rt" event={"ID":"023a246b-31d3-47bb-822b-c8e61aa2f034","Type":"ContainerDied","Data":"59ff3431a84ba005bd2b5c3d863965362c61aebed601f9c5ea3a2b72e47dc82b"} Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.551331 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" event={"ID":"99b291b9-b83d-4854-aa81-3ca6ab88cbee","Type":"ContainerStarted","Data":"ee92d1ec862dd2188275cdb38f6b316b917dd47b902c3f9ca04f6e40fa641854"} Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.551451 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.552517 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt-config-6qk62" event={"ID":"4e24c598-1b77-4133-8840-4e91c8040762","Type":"ContainerDied","Data":"7e1269ed7318deec80f17ebb149b7bc8720dfce0b4d38f763677f498d06f11a1"} Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.552542 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e1269ed7318deec80f17ebb149b7bc8720dfce0b4d38f763677f498d06f11a1" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.552576 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt-config-6qk62" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.591482 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" podStartSLOduration=3.591460561 podStartE2EDuration="3.591460561s" podCreationTimestamp="2025-09-29 19:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:03.586841926 +0000 UTC m=+1065.234631258" watchObservedRunningTime="2025-09-29 19:27:03.591460561 +0000 UTC m=+1065.239249893" Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.623475 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-68snt-config-6qk62"] Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.627997 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-68snt-config-6qk62"] Sep 29 19:27:03 crc kubenswrapper[4741]: I0929 19:27:03.920736 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-68snt" Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.471640 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.840917 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.868468 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-mrkv4"] Sep 29 19:27:04 crc kubenswrapper[4741]: E0929 19:27:04.868806 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e24c598-1b77-4133-8840-4e91c8040762" containerName="ovn-config" Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.868822 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e24c598-1b77-4133-8840-4e91c8040762" containerName="ovn-config" Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.868990 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e24c598-1b77-4133-8840-4e91c8040762" containerName="ovn-config" Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.869515 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mrkv4" Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.883696 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mrkv4"] Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.984048 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-hqxwh"] Sep 29 19:27:04 crc kubenswrapper[4741]: I0929 19:27:04.986677 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hqxwh" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.009638 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hqxwh"] Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.014535 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbws8\" (UniqueName: \"kubernetes.io/projected/4a7cc159-83d9-406f-8e58-23795673b6b0-kube-api-access-hbws8\") pod \"barbican-db-create-mrkv4\" (UID: \"4a7cc159-83d9-406f-8e58-23795673b6b0\") " pod="openstack/barbican-db-create-mrkv4" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.096236 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e24c598-1b77-4133-8840-4e91c8040762" path="/var/lib/kubelet/pods/4e24c598-1b77-4133-8840-4e91c8040762/volumes" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.117143 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnmnr\" (UniqueName: \"kubernetes.io/projected/634aed84-d361-421a-a490-fc9db19afe27-kube-api-access-bnmnr\") pod \"cinder-db-create-hqxwh\" (UID: \"634aed84-d361-421a-a490-fc9db19afe27\") " pod="openstack/cinder-db-create-hqxwh" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.117414 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbws8\" (UniqueName: \"kubernetes.io/projected/4a7cc159-83d9-406f-8e58-23795673b6b0-kube-api-access-hbws8\") pod \"barbican-db-create-mrkv4\" (UID: \"4a7cc159-83d9-406f-8e58-23795673b6b0\") " pod="openstack/barbican-db-create-mrkv4" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.140712 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbws8\" (UniqueName: \"kubernetes.io/projected/4a7cc159-83d9-406f-8e58-23795673b6b0-kube-api-access-hbws8\") pod \"barbican-db-create-mrkv4\" (UID: \"4a7cc159-83d9-406f-8e58-23795673b6b0\") " pod="openstack/barbican-db-create-mrkv4" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.165272 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-7fjg8"] Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.166529 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.171526 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.171638 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.172639 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.180860 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7fjg8"] Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.183796 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kg6nl" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.189698 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mrkv4" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.195570 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-2s8qk"] Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.196751 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2s8qk" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.201976 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z95rt" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.218080 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2s8qk"] Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.218550 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5r84\" (UniqueName: \"kubernetes.io/projected/d38d17bd-4e0b-489d-8b4a-2047a1004e25-kube-api-access-x5r84\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.218596 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-combined-ca-bundle\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.218627 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-config-data\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.218673 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnmnr\" (UniqueName: \"kubernetes.io/projected/634aed84-d361-421a-a490-fc9db19afe27-kube-api-access-bnmnr\") pod \"cinder-db-create-hqxwh\" (UID: \"634aed84-d361-421a-a490-fc9db19afe27\") " pod="openstack/cinder-db-create-hqxwh" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.249520 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnmnr\" (UniqueName: \"kubernetes.io/projected/634aed84-d361-421a-a490-fc9db19afe27-kube-api-access-bnmnr\") pod \"cinder-db-create-hqxwh\" (UID: \"634aed84-d361-421a-a490-fc9db19afe27\") " pod="openstack/cinder-db-create-hqxwh" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.306838 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hqxwh" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.319435 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-combined-ca-bundle\") pod \"023a246b-31d3-47bb-822b-c8e61aa2f034\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.319485 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-db-sync-config-data\") pod \"023a246b-31d3-47bb-822b-c8e61aa2f034\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.319531 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-config-data\") pod \"023a246b-31d3-47bb-822b-c8e61aa2f034\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.319611 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc9b6\" (UniqueName: \"kubernetes.io/projected/023a246b-31d3-47bb-822b-c8e61aa2f034-kube-api-access-vc9b6\") pod \"023a246b-31d3-47bb-822b-c8e61aa2f034\" (UID: \"023a246b-31d3-47bb-822b-c8e61aa2f034\") " Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.319936 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5r84\" (UniqueName: \"kubernetes.io/projected/d38d17bd-4e0b-489d-8b4a-2047a1004e25-kube-api-access-x5r84\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.319960 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-combined-ca-bundle\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.319985 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-config-data\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.320041 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npnnv\" (UniqueName: \"kubernetes.io/projected/c2aad5b5-c32e-4d21-b66c-2b9b39e6341e-kube-api-access-npnnv\") pod \"neutron-db-create-2s8qk\" (UID: \"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e\") " pod="openstack/neutron-db-create-2s8qk" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.324688 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-combined-ca-bundle\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.327287 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-config-data\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.327287 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/023a246b-31d3-47bb-822b-c8e61aa2f034-kube-api-access-vc9b6" (OuterVolumeSpecName: "kube-api-access-vc9b6") pod "023a246b-31d3-47bb-822b-c8e61aa2f034" (UID: "023a246b-31d3-47bb-822b-c8e61aa2f034"). InnerVolumeSpecName "kube-api-access-vc9b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.339035 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "023a246b-31d3-47bb-822b-c8e61aa2f034" (UID: "023a246b-31d3-47bb-822b-c8e61aa2f034"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.347930 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5r84\" (UniqueName: \"kubernetes.io/projected/d38d17bd-4e0b-489d-8b4a-2047a1004e25-kube-api-access-x5r84\") pod \"keystone-db-sync-7fjg8\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.361577 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "023a246b-31d3-47bb-822b-c8e61aa2f034" (UID: "023a246b-31d3-47bb-822b-c8e61aa2f034"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.390375 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-config-data" (OuterVolumeSpecName: "config-data") pod "023a246b-31d3-47bb-822b-c8e61aa2f034" (UID: "023a246b-31d3-47bb-822b-c8e61aa2f034"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.427696 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npnnv\" (UniqueName: \"kubernetes.io/projected/c2aad5b5-c32e-4d21-b66c-2b9b39e6341e-kube-api-access-npnnv\") pod \"neutron-db-create-2s8qk\" (UID: \"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e\") " pod="openstack/neutron-db-create-2s8qk" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.427896 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.427919 4741 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.427933 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/023a246b-31d3-47bb-822b-c8e61aa2f034-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.427947 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc9b6\" (UniqueName: \"kubernetes.io/projected/023a246b-31d3-47bb-822b-c8e61aa2f034-kube-api-access-vc9b6\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.449842 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npnnv\" (UniqueName: \"kubernetes.io/projected/c2aad5b5-c32e-4d21-b66c-2b9b39e6341e-kube-api-access-npnnv\") pod \"neutron-db-create-2s8qk\" (UID: \"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e\") " pod="openstack/neutron-db-create-2s8qk" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.580404 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.587884 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2s8qk" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.631777 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z95rt" event={"ID":"023a246b-31d3-47bb-822b-c8e61aa2f034","Type":"ContainerDied","Data":"9f436e855819643041a51daed5ec88be1ac9c54b7d3d6d76bff7d21c2619e008"} Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.631819 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f436e855819643041a51daed5ec88be1ac9c54b7d3d6d76bff7d21c2619e008" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.631888 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z95rt" Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.678214 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mrkv4"] Sep 29 19:27:05 crc kubenswrapper[4741]: I0929 19:27:05.857838 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hqxwh"] Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.067056 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-m9k4f"] Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.067241 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" podUID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerName="dnsmasq-dns" containerID="cri-o://ee92d1ec862dd2188275cdb38f6b316b917dd47b902c3f9ca04f6e40fa641854" gracePeriod=10 Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.173172 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-77nb7"] Sep 29 19:27:06 crc kubenswrapper[4741]: E0929 19:27:06.173863 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="023a246b-31d3-47bb-822b-c8e61aa2f034" containerName="glance-db-sync" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.173952 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="023a246b-31d3-47bb-822b-c8e61aa2f034" containerName="glance-db-sync" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.176500 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="023a246b-31d3-47bb-822b-c8e61aa2f034" containerName="glance-db-sync" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.177802 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.206787 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-77nb7"] Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.237486 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7fjg8"] Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.253174 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-config\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.253472 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.253619 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.253689 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg2sr\" (UniqueName: \"kubernetes.io/projected/2c0af0d0-0060-4afb-800b-c63b70e69882-kube-api-access-vg2sr\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.253795 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.253875 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: W0929 19:27:06.291710 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd38d17bd_4e0b_489d_8b4a_2047a1004e25.slice/crio-2bd223ae8af4e99cc9f1ef34074ed59336a4ac4b26d9da0cb4a715251ff07c3f WatchSource:0}: Error finding container 2bd223ae8af4e99cc9f1ef34074ed59336a4ac4b26d9da0cb4a715251ff07c3f: Status 404 returned error can't find the container with id 2bd223ae8af4e99cc9f1ef34074ed59336a4ac4b26d9da0cb4a715251ff07c3f Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.296710 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2s8qk"] Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.355888 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.355955 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.355987 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-config\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.356065 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.356140 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.356162 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg2sr\" (UniqueName: \"kubernetes.io/projected/2c0af0d0-0060-4afb-800b-c63b70e69882-kube-api-access-vg2sr\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.358068 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.358618 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.364076 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-config\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.366874 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.372279 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.380689 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg2sr\" (UniqueName: \"kubernetes.io/projected/2c0af0d0-0060-4afb-800b-c63b70e69882-kube-api-access-vg2sr\") pod \"dnsmasq-dns-74f6bcbc87-77nb7\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.550776 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.653049 4741 generic.go:334] "Generic (PLEG): container finished" podID="4a7cc159-83d9-406f-8e58-23795673b6b0" containerID="eb44b7cca7f1c498e95932d205c676cab41a427d9a4d380e41e4ec0ae2e00bb9" exitCode=0 Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.653230 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mrkv4" event={"ID":"4a7cc159-83d9-406f-8e58-23795673b6b0","Type":"ContainerDied","Data":"eb44b7cca7f1c498e95932d205c676cab41a427d9a4d380e41e4ec0ae2e00bb9"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.653376 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mrkv4" event={"ID":"4a7cc159-83d9-406f-8e58-23795673b6b0","Type":"ContainerStarted","Data":"024443c830ba719705779699d82ecba24c92d83b4d7427e9ef69f9c5c68e91b1"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.663811 4741 generic.go:334] "Generic (PLEG): container finished" podID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerID="ee92d1ec862dd2188275cdb38f6b316b917dd47b902c3f9ca04f6e40fa641854" exitCode=0 Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.663828 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" event={"ID":"99b291b9-b83d-4854-aa81-3ca6ab88cbee","Type":"ContainerDied","Data":"ee92d1ec862dd2188275cdb38f6b316b917dd47b902c3f9ca04f6e40fa641854"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.673259 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2s8qk" event={"ID":"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e","Type":"ContainerStarted","Data":"ba8ba5bcbd0b0037bd0c1077997c3f1cb5353af5b057ed27806373a4e5d32b5a"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.673289 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2s8qk" event={"ID":"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e","Type":"ContainerStarted","Data":"825424f765b4e89a368c9f0b8aaf9288312fff1d13b17f6d12199d5920b87cba"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.675555 4741 generic.go:334] "Generic (PLEG): container finished" podID="634aed84-d361-421a-a490-fc9db19afe27" containerID="da0da8b5d3e104bad06c23cfff349af0ee1a80482a6195e6ac53bebbeaf752f2" exitCode=0 Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.675591 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hqxwh" event={"ID":"634aed84-d361-421a-a490-fc9db19afe27","Type":"ContainerDied","Data":"da0da8b5d3e104bad06c23cfff349af0ee1a80482a6195e6ac53bebbeaf752f2"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.675610 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hqxwh" event={"ID":"634aed84-d361-421a-a490-fc9db19afe27","Type":"ContainerStarted","Data":"65f94266f8b7d21539f57f85f6f1a9dccda2c0f178629932181b4dc6d684dd0b"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.677729 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7fjg8" event={"ID":"d38d17bd-4e0b-489d-8b4a-2047a1004e25","Type":"ContainerStarted","Data":"2bd223ae8af4e99cc9f1ef34074ed59336a4ac4b26d9da0cb4a715251ff07c3f"} Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.733342 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-2s8qk" podStartSLOduration=1.733324849 podStartE2EDuration="1.733324849s" podCreationTimestamp="2025-09-29 19:27:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:06.70678938 +0000 UTC m=+1068.354578712" watchObservedRunningTime="2025-09-29 19:27:06.733324849 +0000 UTC m=+1068.381114181" Sep 29 19:27:06 crc kubenswrapper[4741]: I0929 19:27:06.976150 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.068717 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-config\") pod \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.069110 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-swift-storage-0\") pod \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.069142 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-nb\") pod \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.069189 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcfgg\" (UniqueName: \"kubernetes.io/projected/99b291b9-b83d-4854-aa81-3ca6ab88cbee-kube-api-access-lcfgg\") pod \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.069220 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-svc\") pod \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.069252 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-sb\") pod \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\" (UID: \"99b291b9-b83d-4854-aa81-3ca6ab88cbee\") " Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.070548 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-77nb7"] Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.077145 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b291b9-b83d-4854-aa81-3ca6ab88cbee-kube-api-access-lcfgg" (OuterVolumeSpecName: "kube-api-access-lcfgg") pod "99b291b9-b83d-4854-aa81-3ca6ab88cbee" (UID: "99b291b9-b83d-4854-aa81-3ca6ab88cbee"). InnerVolumeSpecName "kube-api-access-lcfgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.143224 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-config" (OuterVolumeSpecName: "config") pod "99b291b9-b83d-4854-aa81-3ca6ab88cbee" (UID: "99b291b9-b83d-4854-aa81-3ca6ab88cbee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.161024 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "99b291b9-b83d-4854-aa81-3ca6ab88cbee" (UID: "99b291b9-b83d-4854-aa81-3ca6ab88cbee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.171271 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.171297 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.171306 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcfgg\" (UniqueName: \"kubernetes.io/projected/99b291b9-b83d-4854-aa81-3ca6ab88cbee-kube-api-access-lcfgg\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.172028 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99b291b9-b83d-4854-aa81-3ca6ab88cbee" (UID: "99b291b9-b83d-4854-aa81-3ca6ab88cbee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.177016 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "99b291b9-b83d-4854-aa81-3ca6ab88cbee" (UID: "99b291b9-b83d-4854-aa81-3ca6ab88cbee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.189116 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "99b291b9-b83d-4854-aa81-3ca6ab88cbee" (UID: "99b291b9-b83d-4854-aa81-3ca6ab88cbee"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.273062 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.273094 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.273103 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99b291b9-b83d-4854-aa81-3ca6ab88cbee-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.570282 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.688799 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" event={"ID":"99b291b9-b83d-4854-aa81-3ca6ab88cbee","Type":"ContainerDied","Data":"5ffc6de5063caa019acd16d2f477e0b11cb997f9a4db647eccf5069959418bd8"} Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.688824 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-m9k4f" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.688862 4741 scope.go:117] "RemoveContainer" containerID="ee92d1ec862dd2188275cdb38f6b316b917dd47b902c3f9ca04f6e40fa641854" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.694096 4741 generic.go:334] "Generic (PLEG): container finished" podID="c2aad5b5-c32e-4d21-b66c-2b9b39e6341e" containerID="ba8ba5bcbd0b0037bd0c1077997c3f1cb5353af5b057ed27806373a4e5d32b5a" exitCode=0 Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.694145 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2s8qk" event={"ID":"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e","Type":"ContainerDied","Data":"ba8ba5bcbd0b0037bd0c1077997c3f1cb5353af5b057ed27806373a4e5d32b5a"} Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.708011 4741 generic.go:334] "Generic (PLEG): container finished" podID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerID="ab5fc5912468e5b08ef1f395eb7eb27bc44632e5a7ac4a11539ac0f08f30db74" exitCode=0 Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.708109 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" event={"ID":"2c0af0d0-0060-4afb-800b-c63b70e69882","Type":"ContainerDied","Data":"ab5fc5912468e5b08ef1f395eb7eb27bc44632e5a7ac4a11539ac0f08f30db74"} Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.708155 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" event={"ID":"2c0af0d0-0060-4afb-800b-c63b70e69882","Type":"ContainerStarted","Data":"c8b093fff128673bcf6e8f30072cbe2ecbcfb4058a0995b2315edf23cf313983"} Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.735763 4741 scope.go:117] "RemoveContainer" containerID="390d29518c9894105621a83295f973d173b7f1e39264f098ed6df915ad191c2b" Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.750243 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-m9k4f"] Sep 29 19:27:07 crc kubenswrapper[4741]: I0929 19:27:07.756358 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-m9k4f"] Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.240281 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hqxwh" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.246202 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mrkv4" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.396842 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbws8\" (UniqueName: \"kubernetes.io/projected/4a7cc159-83d9-406f-8e58-23795673b6b0-kube-api-access-hbws8\") pod \"4a7cc159-83d9-406f-8e58-23795673b6b0\" (UID: \"4a7cc159-83d9-406f-8e58-23795673b6b0\") " Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.396890 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnmnr\" (UniqueName: \"kubernetes.io/projected/634aed84-d361-421a-a490-fc9db19afe27-kube-api-access-bnmnr\") pod \"634aed84-d361-421a-a490-fc9db19afe27\" (UID: \"634aed84-d361-421a-a490-fc9db19afe27\") " Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.404484 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a7cc159-83d9-406f-8e58-23795673b6b0-kube-api-access-hbws8" (OuterVolumeSpecName: "kube-api-access-hbws8") pod "4a7cc159-83d9-406f-8e58-23795673b6b0" (UID: "4a7cc159-83d9-406f-8e58-23795673b6b0"). InnerVolumeSpecName "kube-api-access-hbws8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.404583 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/634aed84-d361-421a-a490-fc9db19afe27-kube-api-access-bnmnr" (OuterVolumeSpecName: "kube-api-access-bnmnr") pod "634aed84-d361-421a-a490-fc9db19afe27" (UID: "634aed84-d361-421a-a490-fc9db19afe27"). InnerVolumeSpecName "kube-api-access-bnmnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.498800 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbws8\" (UniqueName: \"kubernetes.io/projected/4a7cc159-83d9-406f-8e58-23795673b6b0-kube-api-access-hbws8\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.499044 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnmnr\" (UniqueName: \"kubernetes.io/projected/634aed84-d361-421a-a490-fc9db19afe27-kube-api-access-bnmnr\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.716574 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mrkv4" event={"ID":"4a7cc159-83d9-406f-8e58-23795673b6b0","Type":"ContainerDied","Data":"024443c830ba719705779699d82ecba24c92d83b4d7427e9ef69f9c5c68e91b1"} Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.716611 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="024443c830ba719705779699d82ecba24c92d83b4d7427e9ef69f9c5c68e91b1" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.716673 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mrkv4" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.722096 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" event={"ID":"2c0af0d0-0060-4afb-800b-c63b70e69882","Type":"ContainerStarted","Data":"8866fd403a8f411ea9d2d824752f67513dcb07bc65ebf9422999d845c9b65f48"} Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.722158 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.723891 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hqxwh" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.723883 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hqxwh" event={"ID":"634aed84-d361-421a-a490-fc9db19afe27","Type":"ContainerDied","Data":"65f94266f8b7d21539f57f85f6f1a9dccda2c0f178629932181b4dc6d684dd0b"} Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.723921 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65f94266f8b7d21539f57f85f6f1a9dccda2c0f178629932181b4dc6d684dd0b" Sep 29 19:27:08 crc kubenswrapper[4741]: I0929 19:27:08.754200 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" podStartSLOduration=2.754182015 podStartE2EDuration="2.754182015s" podCreationTimestamp="2025-09-29 19:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:08.748562067 +0000 UTC m=+1070.396351419" watchObservedRunningTime="2025-09-29 19:27:08.754182015 +0000 UTC m=+1070.401971347" Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.127207 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" path="/var/lib/kubelet/pods/99b291b9-b83d-4854-aa81-3ca6ab88cbee/volumes" Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.210504 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2s8qk" Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.312734 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npnnv\" (UniqueName: \"kubernetes.io/projected/c2aad5b5-c32e-4d21-b66c-2b9b39e6341e-kube-api-access-npnnv\") pod \"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e\" (UID: \"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e\") " Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.320358 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2aad5b5-c32e-4d21-b66c-2b9b39e6341e-kube-api-access-npnnv" (OuterVolumeSpecName: "kube-api-access-npnnv") pod "c2aad5b5-c32e-4d21-b66c-2b9b39e6341e" (UID: "c2aad5b5-c32e-4d21-b66c-2b9b39e6341e"). InnerVolumeSpecName "kube-api-access-npnnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.414839 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npnnv\" (UniqueName: \"kubernetes.io/projected/c2aad5b5-c32e-4d21-b66c-2b9b39e6341e-kube-api-access-npnnv\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.732632 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2s8qk" Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.732632 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2s8qk" event={"ID":"c2aad5b5-c32e-4d21-b66c-2b9b39e6341e","Type":"ContainerDied","Data":"825424f765b4e89a368c9f0b8aaf9288312fff1d13b17f6d12199d5920b87cba"} Sep 29 19:27:13 crc kubenswrapper[4741]: I0929 19:27:09.732674 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="825424f765b4e89a368c9f0b8aaf9288312fff1d13b17f6d12199d5920b87cba" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.902542 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-b41f-account-create-nsqlq"] Sep 29 19:27:14 crc kubenswrapper[4741]: E0929 19:27:14.903352 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2aad5b5-c32e-4d21-b66c-2b9b39e6341e" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903363 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2aad5b5-c32e-4d21-b66c-2b9b39e6341e" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: E0929 19:27:14.903375 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerName="init" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903381 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerName="init" Sep 29 19:27:14 crc kubenswrapper[4741]: E0929 19:27:14.903414 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerName="dnsmasq-dns" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903420 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerName="dnsmasq-dns" Sep 29 19:27:14 crc kubenswrapper[4741]: E0929 19:27:14.903431 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7cc159-83d9-406f-8e58-23795673b6b0" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903437 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7cc159-83d9-406f-8e58-23795673b6b0" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: E0929 19:27:14.903456 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634aed84-d361-421a-a490-fc9db19afe27" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903461 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="634aed84-d361-421a-a490-fc9db19afe27" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903616 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="634aed84-d361-421a-a490-fc9db19afe27" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903629 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a7cc159-83d9-406f-8e58-23795673b6b0" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903648 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b291b9-b83d-4854-aa81-3ca6ab88cbee" containerName="dnsmasq-dns" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.903657 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2aad5b5-c32e-4d21-b66c-2b9b39e6341e" containerName="mariadb-database-create" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.904127 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b41f-account-create-nsqlq" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.909139 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Sep 29 19:27:14 crc kubenswrapper[4741]: I0929 19:27:14.915094 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b41f-account-create-nsqlq"] Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.001598 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e788-account-create-f9d6m"] Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.003182 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e788-account-create-f9d6m" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.007002 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.014990 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e788-account-create-f9d6m"] Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.018070 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84j2c\" (UniqueName: \"kubernetes.io/projected/9a3f71de-d7bd-4c9c-b44c-63016e612edc-kube-api-access-84j2c\") pod \"cinder-b41f-account-create-nsqlq\" (UID: \"9a3f71de-d7bd-4c9c-b44c-63016e612edc\") " pod="openstack/cinder-b41f-account-create-nsqlq" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.119799 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84j2c\" (UniqueName: \"kubernetes.io/projected/9a3f71de-d7bd-4c9c-b44c-63016e612edc-kube-api-access-84j2c\") pod \"cinder-b41f-account-create-nsqlq\" (UID: \"9a3f71de-d7bd-4c9c-b44c-63016e612edc\") " pod="openstack/cinder-b41f-account-create-nsqlq" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.119898 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5b74\" (UniqueName: \"kubernetes.io/projected/91807a37-17a2-451a-bc6a-69f7cd09707c-kube-api-access-g5b74\") pod \"barbican-e788-account-create-f9d6m\" (UID: \"91807a37-17a2-451a-bc6a-69f7cd09707c\") " pod="openstack/barbican-e788-account-create-f9d6m" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.138348 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84j2c\" (UniqueName: \"kubernetes.io/projected/9a3f71de-d7bd-4c9c-b44c-63016e612edc-kube-api-access-84j2c\") pod \"cinder-b41f-account-create-nsqlq\" (UID: \"9a3f71de-d7bd-4c9c-b44c-63016e612edc\") " pod="openstack/cinder-b41f-account-create-nsqlq" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.206763 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-07da-account-create-wcv7g"] Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.207884 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-07da-account-create-wcv7g" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.213042 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.218260 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-07da-account-create-wcv7g"] Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.230887 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b41f-account-create-nsqlq" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.231411 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5b74\" (UniqueName: \"kubernetes.io/projected/91807a37-17a2-451a-bc6a-69f7cd09707c-kube-api-access-g5b74\") pod \"barbican-e788-account-create-f9d6m\" (UID: \"91807a37-17a2-451a-bc6a-69f7cd09707c\") " pod="openstack/barbican-e788-account-create-f9d6m" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.251156 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5b74\" (UniqueName: \"kubernetes.io/projected/91807a37-17a2-451a-bc6a-69f7cd09707c-kube-api-access-g5b74\") pod \"barbican-e788-account-create-f9d6m\" (UID: \"91807a37-17a2-451a-bc6a-69f7cd09707c\") " pod="openstack/barbican-e788-account-create-f9d6m" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.324568 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e788-account-create-f9d6m" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.333082 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54v55\" (UniqueName: \"kubernetes.io/projected/21479272-e739-4a74-a149-8c63a5800975-kube-api-access-54v55\") pod \"neutron-07da-account-create-wcv7g\" (UID: \"21479272-e739-4a74-a149-8c63a5800975\") " pod="openstack/neutron-07da-account-create-wcv7g" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.435148 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54v55\" (UniqueName: \"kubernetes.io/projected/21479272-e739-4a74-a149-8c63a5800975-kube-api-access-54v55\") pod \"neutron-07da-account-create-wcv7g\" (UID: \"21479272-e739-4a74-a149-8c63a5800975\") " pod="openstack/neutron-07da-account-create-wcv7g" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.456538 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54v55\" (UniqueName: \"kubernetes.io/projected/21479272-e739-4a74-a149-8c63a5800975-kube-api-access-54v55\") pod \"neutron-07da-account-create-wcv7g\" (UID: \"21479272-e739-4a74-a149-8c63a5800975\") " pod="openstack/neutron-07da-account-create-wcv7g" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.652360 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b41f-account-create-nsqlq"] Sep 29 19:27:15 crc kubenswrapper[4741]: W0929 19:27:15.654018 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a3f71de_d7bd_4c9c_b44c_63016e612edc.slice/crio-e2eebff40095184d66cc25a39fb7abba005ca638305431b28db38896593816dc WatchSource:0}: Error finding container e2eebff40095184d66cc25a39fb7abba005ca638305431b28db38896593816dc: Status 404 returned error can't find the container with id e2eebff40095184d66cc25a39fb7abba005ca638305431b28db38896593816dc Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.723451 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-07da-account-create-wcv7g" Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.795492 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e788-account-create-f9d6m"] Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.799462 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b41f-account-create-nsqlq" event={"ID":"9a3f71de-d7bd-4c9c-b44c-63016e612edc","Type":"ContainerStarted","Data":"e2eebff40095184d66cc25a39fb7abba005ca638305431b28db38896593816dc"} Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.807964 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7fjg8" event={"ID":"d38d17bd-4e0b-489d-8b4a-2047a1004e25","Type":"ContainerStarted","Data":"acd32fd79bc7d0c38d24d970d07163f84e2a2deb7951623bb290c323bb8d913a"} Sep 29 19:27:15 crc kubenswrapper[4741]: W0929 19:27:15.809147 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91807a37_17a2_451a_bc6a_69f7cd09707c.slice/crio-377fa9bc9f3d7b4704f82410b70c055db6d95c391312d92d7450f678c0bed4f3 WatchSource:0}: Error finding container 377fa9bc9f3d7b4704f82410b70c055db6d95c391312d92d7450f678c0bed4f3: Status 404 returned error can't find the container with id 377fa9bc9f3d7b4704f82410b70c055db6d95c391312d92d7450f678c0bed4f3 Sep 29 19:27:15 crc kubenswrapper[4741]: I0929 19:27:15.831771 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-7fjg8" podStartSLOduration=2.033050757 podStartE2EDuration="10.831750363s" podCreationTimestamp="2025-09-29 19:27:05 +0000 UTC" firstStartedPulling="2025-09-29 19:27:06.33205689 +0000 UTC m=+1067.979846222" lastFinishedPulling="2025-09-29 19:27:15.130756496 +0000 UTC m=+1076.778545828" observedRunningTime="2025-09-29 19:27:15.826521018 +0000 UTC m=+1077.474310380" watchObservedRunningTime="2025-09-29 19:27:15.831750363 +0000 UTC m=+1077.479539695" Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.197148 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-07da-account-create-wcv7g"] Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.552429 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.607710 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sxlh8"] Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.607988 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-sxlh8" podUID="07649762-f417-4bfe-877a-1526d40f2498" containerName="dnsmasq-dns" containerID="cri-o://5da6e4afbcf518eac9d8729284ae631a8a8d08c43a7f4a87fd06e085bdd83378" gracePeriod=10 Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.824775 4741 generic.go:334] "Generic (PLEG): container finished" podID="9a3f71de-d7bd-4c9c-b44c-63016e612edc" containerID="f996a2898035c79af3529c1361930f4c58e64aae7f679f404e92d59c0e8200c0" exitCode=0 Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.824864 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b41f-account-create-nsqlq" event={"ID":"9a3f71de-d7bd-4c9c-b44c-63016e612edc","Type":"ContainerDied","Data":"f996a2898035c79af3529c1361930f4c58e64aae7f679f404e92d59c0e8200c0"} Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.828223 4741 generic.go:334] "Generic (PLEG): container finished" podID="07649762-f417-4bfe-877a-1526d40f2498" containerID="5da6e4afbcf518eac9d8729284ae631a8a8d08c43a7f4a87fd06e085bdd83378" exitCode=0 Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.828259 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sxlh8" event={"ID":"07649762-f417-4bfe-877a-1526d40f2498","Type":"ContainerDied","Data":"5da6e4afbcf518eac9d8729284ae631a8a8d08c43a7f4a87fd06e085bdd83378"} Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.830008 4741 generic.go:334] "Generic (PLEG): container finished" podID="21479272-e739-4a74-a149-8c63a5800975" containerID="c840b9d34193b3248a4a42d0dfa2c7dcbcf07b30ff0be4e77faf1f7aa259dda9" exitCode=0 Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.830068 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-07da-account-create-wcv7g" event={"ID":"21479272-e739-4a74-a149-8c63a5800975","Type":"ContainerDied","Data":"c840b9d34193b3248a4a42d0dfa2c7dcbcf07b30ff0be4e77faf1f7aa259dda9"} Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.830085 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-07da-account-create-wcv7g" event={"ID":"21479272-e739-4a74-a149-8c63a5800975","Type":"ContainerStarted","Data":"594bc4da249ad42b6ad7480bef1321b0b4fb17a38974351ffc53d3835b1e89c3"} Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.831627 4741 generic.go:334] "Generic (PLEG): container finished" podID="91807a37-17a2-451a-bc6a-69f7cd09707c" containerID="f24cdfed73077cdb3c2d36e877e66cfcc8b22dcefabe794c04d88c2e3b0f8099" exitCode=0 Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.831706 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e788-account-create-f9d6m" event={"ID":"91807a37-17a2-451a-bc6a-69f7cd09707c","Type":"ContainerDied","Data":"f24cdfed73077cdb3c2d36e877e66cfcc8b22dcefabe794c04d88c2e3b0f8099"} Sep 29 19:27:16 crc kubenswrapper[4741]: I0929 19:27:16.831730 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e788-account-create-f9d6m" event={"ID":"91807a37-17a2-451a-bc6a-69f7cd09707c","Type":"ContainerStarted","Data":"377fa9bc9f3d7b4704f82410b70c055db6d95c391312d92d7450f678c0bed4f3"} Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.102427 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.286344 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-nb\") pod \"07649762-f417-4bfe-877a-1526d40f2498\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.286497 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-config\") pod \"07649762-f417-4bfe-877a-1526d40f2498\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.286549 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-sb\") pod \"07649762-f417-4bfe-877a-1526d40f2498\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.286629 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkn9w\" (UniqueName: \"kubernetes.io/projected/07649762-f417-4bfe-877a-1526d40f2498-kube-api-access-gkn9w\") pod \"07649762-f417-4bfe-877a-1526d40f2498\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.286690 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-dns-svc\") pod \"07649762-f417-4bfe-877a-1526d40f2498\" (UID: \"07649762-f417-4bfe-877a-1526d40f2498\") " Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.294446 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07649762-f417-4bfe-877a-1526d40f2498-kube-api-access-gkn9w" (OuterVolumeSpecName: "kube-api-access-gkn9w") pod "07649762-f417-4bfe-877a-1526d40f2498" (UID: "07649762-f417-4bfe-877a-1526d40f2498"). InnerVolumeSpecName "kube-api-access-gkn9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.337249 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "07649762-f417-4bfe-877a-1526d40f2498" (UID: "07649762-f417-4bfe-877a-1526d40f2498"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.343816 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-config" (OuterVolumeSpecName: "config") pod "07649762-f417-4bfe-877a-1526d40f2498" (UID: "07649762-f417-4bfe-877a-1526d40f2498"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.352901 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "07649762-f417-4bfe-877a-1526d40f2498" (UID: "07649762-f417-4bfe-877a-1526d40f2498"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.355291 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "07649762-f417-4bfe-877a-1526d40f2498" (UID: "07649762-f417-4bfe-877a-1526d40f2498"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.388590 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.388622 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.388633 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.388642 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07649762-f417-4bfe-877a-1526d40f2498-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.388651 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkn9w\" (UniqueName: \"kubernetes.io/projected/07649762-f417-4bfe-877a-1526d40f2498-kube-api-access-gkn9w\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.845481 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sxlh8" event={"ID":"07649762-f417-4bfe-877a-1526d40f2498","Type":"ContainerDied","Data":"8d5b08d3c1d11d655c028172403fe99c4d8d41b8284f8e04d6146ea25ef4cec1"} Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.845565 4741 scope.go:117] "RemoveContainer" containerID="5da6e4afbcf518eac9d8729284ae631a8a8d08c43a7f4a87fd06e085bdd83378" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.845961 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sxlh8" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.869399 4741 scope.go:117] "RemoveContainer" containerID="8dcb2e8e8e114666df2ee7d5a78448ed7c3968bf7de7f2c22e93a9682facb037" Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.900637 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sxlh8"] Sep 29 19:27:17 crc kubenswrapper[4741]: I0929 19:27:17.909496 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sxlh8"] Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.236973 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e788-account-create-f9d6m" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.328441 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b41f-account-create-nsqlq" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.334326 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-07da-account-create-wcv7g" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.405537 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5b74\" (UniqueName: \"kubernetes.io/projected/91807a37-17a2-451a-bc6a-69f7cd09707c-kube-api-access-g5b74\") pod \"91807a37-17a2-451a-bc6a-69f7cd09707c\" (UID: \"91807a37-17a2-451a-bc6a-69f7cd09707c\") " Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.418522 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91807a37-17a2-451a-bc6a-69f7cd09707c-kube-api-access-g5b74" (OuterVolumeSpecName: "kube-api-access-g5b74") pod "91807a37-17a2-451a-bc6a-69f7cd09707c" (UID: "91807a37-17a2-451a-bc6a-69f7cd09707c"). InnerVolumeSpecName "kube-api-access-g5b74". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.506559 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54v55\" (UniqueName: \"kubernetes.io/projected/21479272-e739-4a74-a149-8c63a5800975-kube-api-access-54v55\") pod \"21479272-e739-4a74-a149-8c63a5800975\" (UID: \"21479272-e739-4a74-a149-8c63a5800975\") " Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.506655 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84j2c\" (UniqueName: \"kubernetes.io/projected/9a3f71de-d7bd-4c9c-b44c-63016e612edc-kube-api-access-84j2c\") pod \"9a3f71de-d7bd-4c9c-b44c-63016e612edc\" (UID: \"9a3f71de-d7bd-4c9c-b44c-63016e612edc\") " Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.507003 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5b74\" (UniqueName: \"kubernetes.io/projected/91807a37-17a2-451a-bc6a-69f7cd09707c-kube-api-access-g5b74\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.509695 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3f71de-d7bd-4c9c-b44c-63016e612edc-kube-api-access-84j2c" (OuterVolumeSpecName: "kube-api-access-84j2c") pod "9a3f71de-d7bd-4c9c-b44c-63016e612edc" (UID: "9a3f71de-d7bd-4c9c-b44c-63016e612edc"). InnerVolumeSpecName "kube-api-access-84j2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.509809 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21479272-e739-4a74-a149-8c63a5800975-kube-api-access-54v55" (OuterVolumeSpecName: "kube-api-access-54v55") pod "21479272-e739-4a74-a149-8c63a5800975" (UID: "21479272-e739-4a74-a149-8c63a5800975"). InnerVolumeSpecName "kube-api-access-54v55". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.608873 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54v55\" (UniqueName: \"kubernetes.io/projected/21479272-e739-4a74-a149-8c63a5800975-kube-api-access-54v55\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.608936 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84j2c\" (UniqueName: \"kubernetes.io/projected/9a3f71de-d7bd-4c9c-b44c-63016e612edc-kube-api-access-84j2c\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.872953 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-07da-account-create-wcv7g" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.873429 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-07da-account-create-wcv7g" event={"ID":"21479272-e739-4a74-a149-8c63a5800975","Type":"ContainerDied","Data":"594bc4da249ad42b6ad7480bef1321b0b4fb17a38974351ffc53d3835b1e89c3"} Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.873521 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="594bc4da249ad42b6ad7480bef1321b0b4fb17a38974351ffc53d3835b1e89c3" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.876092 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e788-account-create-f9d6m" event={"ID":"91807a37-17a2-451a-bc6a-69f7cd09707c","Type":"ContainerDied","Data":"377fa9bc9f3d7b4704f82410b70c055db6d95c391312d92d7450f678c0bed4f3"} Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.876114 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e788-account-create-f9d6m" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.876128 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="377fa9bc9f3d7b4704f82410b70c055db6d95c391312d92d7450f678c0bed4f3" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.882948 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b41f-account-create-nsqlq" event={"ID":"9a3f71de-d7bd-4c9c-b44c-63016e612edc","Type":"ContainerDied","Data":"e2eebff40095184d66cc25a39fb7abba005ca638305431b28db38896593816dc"} Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.883030 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2eebff40095184d66cc25a39fb7abba005ca638305431b28db38896593816dc" Sep 29 19:27:18 crc kubenswrapper[4741]: I0929 19:27:18.883094 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b41f-account-create-nsqlq" Sep 29 19:27:19 crc kubenswrapper[4741]: I0929 19:27:19.103502 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07649762-f417-4bfe-877a-1526d40f2498" path="/var/lib/kubelet/pods/07649762-f417-4bfe-877a-1526d40f2498/volumes" Sep 29 19:27:19 crc kubenswrapper[4741]: I0929 19:27:19.890975 4741 generic.go:334] "Generic (PLEG): container finished" podID="d38d17bd-4e0b-489d-8b4a-2047a1004e25" containerID="acd32fd79bc7d0c38d24d970d07163f84e2a2deb7951623bb290c323bb8d913a" exitCode=0 Sep 29 19:27:19 crc kubenswrapper[4741]: I0929 19:27:19.891152 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7fjg8" event={"ID":"d38d17bd-4e0b-489d-8b4a-2047a1004e25","Type":"ContainerDied","Data":"acd32fd79bc7d0c38d24d970d07163f84e2a2deb7951623bb290c323bb8d913a"} Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.235760 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.351270 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5r84\" (UniqueName: \"kubernetes.io/projected/d38d17bd-4e0b-489d-8b4a-2047a1004e25-kube-api-access-x5r84\") pod \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.351504 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-combined-ca-bundle\") pod \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.351660 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-config-data\") pod \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\" (UID: \"d38d17bd-4e0b-489d-8b4a-2047a1004e25\") " Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.364131 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38d17bd-4e0b-489d-8b4a-2047a1004e25-kube-api-access-x5r84" (OuterVolumeSpecName: "kube-api-access-x5r84") pod "d38d17bd-4e0b-489d-8b4a-2047a1004e25" (UID: "d38d17bd-4e0b-489d-8b4a-2047a1004e25"). InnerVolumeSpecName "kube-api-access-x5r84". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.372676 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d38d17bd-4e0b-489d-8b4a-2047a1004e25" (UID: "d38d17bd-4e0b-489d-8b4a-2047a1004e25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.388908 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-config-data" (OuterVolumeSpecName: "config-data") pod "d38d17bd-4e0b-489d-8b4a-2047a1004e25" (UID: "d38d17bd-4e0b-489d-8b4a-2047a1004e25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.453761 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.453793 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5r84\" (UniqueName: \"kubernetes.io/projected/d38d17bd-4e0b-489d-8b4a-2047a1004e25-kube-api-access-x5r84\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.453804 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38d17bd-4e0b-489d-8b4a-2047a1004e25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.917215 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7fjg8" event={"ID":"d38d17bd-4e0b-489d-8b4a-2047a1004e25","Type":"ContainerDied","Data":"2bd223ae8af4e99cc9f1ef34074ed59336a4ac4b26d9da0cb4a715251ff07c3f"} Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.917257 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bd223ae8af4e99cc9f1ef34074ed59336a4ac4b26d9da0cb4a715251ff07c3f" Sep 29 19:27:21 crc kubenswrapper[4741]: I0929 19:27:21.917322 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7fjg8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.184190 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-xjq8s"] Sep 29 19:27:22 crc kubenswrapper[4741]: E0929 19:27:22.184880 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07649762-f417-4bfe-877a-1526d40f2498" containerName="init" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.184901 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="07649762-f417-4bfe-877a-1526d40f2498" containerName="init" Sep 29 19:27:22 crc kubenswrapper[4741]: E0929 19:27:22.184926 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21479272-e739-4a74-a149-8c63a5800975" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.184935 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="21479272-e739-4a74-a149-8c63a5800975" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: E0929 19:27:22.184964 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3f71de-d7bd-4c9c-b44c-63016e612edc" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.184973 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3f71de-d7bd-4c9c-b44c-63016e612edc" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: E0929 19:27:22.184985 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07649762-f417-4bfe-877a-1526d40f2498" containerName="dnsmasq-dns" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.184993 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="07649762-f417-4bfe-877a-1526d40f2498" containerName="dnsmasq-dns" Sep 29 19:27:22 crc kubenswrapper[4741]: E0929 19:27:22.185011 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38d17bd-4e0b-489d-8b4a-2047a1004e25" containerName="keystone-db-sync" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.185019 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38d17bd-4e0b-489d-8b4a-2047a1004e25" containerName="keystone-db-sync" Sep 29 19:27:22 crc kubenswrapper[4741]: E0929 19:27:22.185032 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91807a37-17a2-451a-bc6a-69f7cd09707c" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.185041 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="91807a37-17a2-451a-bc6a-69f7cd09707c" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.185228 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3f71de-d7bd-4c9c-b44c-63016e612edc" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.185248 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="91807a37-17a2-451a-bc6a-69f7cd09707c" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.185264 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="21479272-e739-4a74-a149-8c63a5800975" containerName="mariadb-account-create" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.185287 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="07649762-f417-4bfe-877a-1526d40f2498" containerName="dnsmasq-dns" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.185300 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38d17bd-4e0b-489d-8b4a-2047a1004e25" containerName="keystone-db-sync" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.186381 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.202145 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rv4kz"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.203264 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.209029 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-xjq8s"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.211886 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.212225 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.212426 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.212620 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kg6nl" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.228708 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rv4kz"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267124 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-combined-ca-bundle\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267199 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-credential-keys\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267228 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-svc\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267283 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs8v9\" (UniqueName: \"kubernetes.io/projected/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-kube-api-access-bs8v9\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267322 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-config\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267352 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267367 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-fernet-keys\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267413 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkb4b\" (UniqueName: \"kubernetes.io/projected/6e632b91-9db1-4c5b-adb4-921af2dbcabe-kube-api-access-kkb4b\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267461 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-scripts\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267493 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267511 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-config-data\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.267530 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369058 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-scripts\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369102 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369123 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-config-data\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369145 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369165 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-combined-ca-bundle\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369203 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-credential-keys\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369228 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-svc\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369268 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs8v9\" (UniqueName: \"kubernetes.io/projected/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-kube-api-access-bs8v9\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369286 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-config\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369316 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369332 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-fernet-keys\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.369349 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkb4b\" (UniqueName: \"kubernetes.io/projected/6e632b91-9db1-4c5b-adb4-921af2dbcabe-kube-api-access-kkb4b\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.370483 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-svc\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.370605 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.371153 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.371185 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-config\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.372078 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.374357 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-scripts\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.375233 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-combined-ca-bundle\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.375368 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-config-data\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.379948 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-fernet-keys\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.381907 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-credential-keys\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.414816 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-xlz27"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.424842 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.450751 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.450871 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-c9t62" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.451020 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.466702 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs8v9\" (UniqueName: \"kubernetes.io/projected/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-kube-api-access-bs8v9\") pod \"keystone-bootstrap-rv4kz\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.472093 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkb4b\" (UniqueName: \"kubernetes.io/projected/6e632b91-9db1-4c5b-adb4-921af2dbcabe-kube-api-access-kkb4b\") pod \"dnsmasq-dns-847c4cc679-xjq8s\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.516815 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.526455 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-gfpz8"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.527756 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.534197 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.537527 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.550907 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.551273 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.551407 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-vb8vq" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.551539 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.556996 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.563669 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gfpz8"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.594902 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.598973 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-config\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599362 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-combined-ca-bundle\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599410 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc4nn\" (UniqueName: \"kubernetes.io/projected/72e81d61-d517-412d-a5cb-100a5aac1ec3-kube-api-access-sc4nn\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599436 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-db-sync-config-data\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599476 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e172bee5-8543-4044-bf71-57e488d14fe4-etc-machine-id\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599499 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-combined-ca-bundle\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599525 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-config-data\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599556 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-scripts\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.599578 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67s8w\" (UniqueName: \"kubernetes.io/projected/e172bee5-8543-4044-bf71-57e488d14fe4-kube-api-access-67s8w\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.640764 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.664454 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xlz27"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703257 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e172bee5-8543-4044-bf71-57e488d14fe4-etc-machine-id\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703310 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-combined-ca-bundle\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703342 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-config-data\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703409 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703457 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-scripts\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703490 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67s8w\" (UniqueName: \"kubernetes.io/projected/e172bee5-8543-4044-bf71-57e488d14fe4-kube-api-access-67s8w\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703511 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-run-httpd\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703548 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-config\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703572 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxjtt\" (UniqueName: \"kubernetes.io/projected/34dc5bf3-42dd-47ec-ba42-7422df4dde51-kube-api-access-zxjtt\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703628 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-combined-ca-bundle\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703655 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-scripts\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703675 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc4nn\" (UniqueName: \"kubernetes.io/projected/72e81d61-d517-412d-a5cb-100a5aac1ec3-kube-api-access-sc4nn\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703689 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-log-httpd\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703707 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-config-data\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703729 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.703747 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-db-sync-config-data\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.713533 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-db-sync-config-data\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.717812 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e172bee5-8543-4044-bf71-57e488d14fe4-etc-machine-id\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.723553 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-config\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.743171 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-combined-ca-bundle\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.743621 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-config-data\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.753057 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-combined-ca-bundle\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.758087 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67s8w\" (UniqueName: \"kubernetes.io/projected/e172bee5-8543-4044-bf71-57e488d14fe4-kube-api-access-67s8w\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.764215 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc4nn\" (UniqueName: \"kubernetes.io/projected/72e81d61-d517-412d-a5cb-100a5aac1ec3-kube-api-access-sc4nn\") pod \"neutron-db-sync-xlz27\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.764889 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-scripts\") pod \"cinder-db-sync-gfpz8\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.800457 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hjb7q"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.801551 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.805969 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.806362 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.806478 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-xjq8s"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.814817 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pwnw4" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.815823 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-run-httpd\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.815860 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxjtt\" (UniqueName: \"kubernetes.io/projected/34dc5bf3-42dd-47ec-ba42-7422df4dde51-kube-api-access-zxjtt\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.815920 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-scripts\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.815939 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-log-httpd\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.815957 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-config-data\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.815974 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.816019 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.816463 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-log-httpd\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.816658 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-run-httpd\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.834303 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-scripts\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.841002 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hjb7q"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.851715 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.854664 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-646p8"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.855764 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.862773 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.862802 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x729m" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.863355 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.863872 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-646p8"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.867844 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxjtt\" (UniqueName: \"kubernetes.io/projected/34dc5bf3-42dd-47ec-ba42-7422df4dde51-kube-api-access-zxjtt\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.868301 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-config-data\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.868562 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " pod="openstack/ceilometer-0" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.873664 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-nt54p"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.875488 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.891928 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-nt54p"] Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.917163 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-config-data\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.917203 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13ca2f00-5599-4912-83dd-afbe369673b6-logs\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.917302 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-combined-ca-bundle\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.917512 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fss9\" (UniqueName: \"kubernetes.io/projected/13ca2f00-5599-4912-83dd-afbe369673b6-kube-api-access-5fss9\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.917571 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-scripts\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.924927 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:27:22 crc kubenswrapper[4741]: I0929 19:27:22.946915 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020260 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-db-sync-config-data\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020531 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-config-data\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020557 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13ca2f00-5599-4912-83dd-afbe369673b6-logs\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020600 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-combined-ca-bundle\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020679 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020719 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020769 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-combined-ca-bundle\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020817 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sglhw\" (UniqueName: \"kubernetes.io/projected/00111435-439e-47b2-b8e8-08ee8fa40444-kube-api-access-sglhw\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020848 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-config\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020897 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fss9\" (UniqueName: \"kubernetes.io/projected/13ca2f00-5599-4912-83dd-afbe369673b6-kube-api-access-5fss9\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020920 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020935 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-scripts\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.020995 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5fg7\" (UniqueName: \"kubernetes.io/projected/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-kube-api-access-q5fg7\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.021142 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.025331 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13ca2f00-5599-4912-83dd-afbe369673b6-logs\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.035230 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-config-data\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.035533 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-scripts\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.037122 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-combined-ca-bundle\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.045654 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fss9\" (UniqueName: \"kubernetes.io/projected/13ca2f00-5599-4912-83dd-afbe369673b6-kube-api-access-5fss9\") pod \"placement-db-sync-hjb7q\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122265 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122319 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sglhw\" (UniqueName: \"kubernetes.io/projected/00111435-439e-47b2-b8e8-08ee8fa40444-kube-api-access-sglhw\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122342 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-config\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122369 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122415 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5fg7\" (UniqueName: \"kubernetes.io/projected/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-kube-api-access-q5fg7\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122437 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122459 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-db-sync-config-data\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122485 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-combined-ca-bundle\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.122522 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.123753 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.123872 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.125727 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.126884 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-config\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.127705 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.131143 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-db-sync-config-data\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.143694 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sglhw\" (UniqueName: \"kubernetes.io/projected/00111435-439e-47b2-b8e8-08ee8fa40444-kube-api-access-sglhw\") pod \"dnsmasq-dns-785d8bcb8c-nt54p\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.145981 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-combined-ca-bundle\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.156125 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5fg7\" (UniqueName: \"kubernetes.io/projected/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-kube-api-access-q5fg7\") pod \"barbican-db-sync-646p8\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.195068 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.242245 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-xjq8s"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.242593 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-646p8" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.256805 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.310666 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.330985 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.333156 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-h9mn9" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.333417 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.333456 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.335125 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.335744 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.381641 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.401319 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.407295 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.407643 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431116 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431193 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-logs\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431298 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431345 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-scripts\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431371 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-config-data\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431402 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431924 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.431957 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpzz4\" (UniqueName: \"kubernetes.io/projected/88ace222-b0cb-492d-a0c8-3421e59a074e-kube-api-access-qpzz4\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.462085 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.521287 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rv4kz"] Sep 29 19:27:23 crc kubenswrapper[4741]: W0929 19:27:23.528759 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode172bee5_8543_4044_bf71_57e488d14fe4.slice/crio-4ed66f64565d1f1c11d4dc81367805d6d8aa89947eef157cd3134336fcb5aea7 WatchSource:0}: Error finding container 4ed66f64565d1f1c11d4dc81367805d6d8aa89947eef157cd3134336fcb5aea7: Status 404 returned error can't find the container with id 4ed66f64565d1f1c11d4dc81367805d6d8aa89947eef157cd3134336fcb5aea7 Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533241 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-scripts\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-config-data\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533313 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lhqs\" (UniqueName: \"kubernetes.io/projected/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-kube-api-access-4lhqs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533336 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533370 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533403 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpzz4\" (UniqueName: \"kubernetes.io/projected/88ace222-b0cb-492d-a0c8-3421e59a074e-kube-api-access-qpzz4\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533424 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533457 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-logs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533480 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533499 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533514 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533542 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-logs\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533559 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533608 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533645 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.533685 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.534027 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.535271 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.539074 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-logs\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.540549 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-scripts\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.540827 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.558291 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gfpz8"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.561591 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.563773 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-config-data\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.574167 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.577044 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpzz4\" (UniqueName: \"kubernetes.io/projected/88ace222-b0cb-492d-a0c8-3421e59a074e-kube-api-access-qpzz4\") pod \"glance-default-external-api-0\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.578515 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xlz27"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637374 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637458 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637518 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lhqs\" (UniqueName: \"kubernetes.io/projected/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-kube-api-access-4lhqs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637575 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637594 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-logs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637613 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637629 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.637655 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.641110 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.641558 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.641795 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-logs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.641983 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.644694 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.653565 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.654492 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.661093 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lhqs\" (UniqueName: \"kubernetes.io/projected/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-kube-api-access-4lhqs\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.671746 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.706190 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.716534 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.879469 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-nt54p"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.897034 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-646p8"] Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.903702 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hjb7q"] Sep 29 19:27:23 crc kubenswrapper[4741]: W0929 19:27:23.923277 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e91d309_4c23_4d57_ad3f_8eb7634b9b04.slice/crio-2556098fc6e89af7c6c351ad10e0573f555efe8c133326fa21eefb11b9902401 WatchSource:0}: Error finding container 2556098fc6e89af7c6c351ad10e0573f555efe8c133326fa21eefb11b9902401: Status 404 returned error can't find the container with id 2556098fc6e89af7c6c351ad10e0573f555efe8c133326fa21eefb11b9902401 Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.958665 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hjb7q" event={"ID":"13ca2f00-5599-4912-83dd-afbe369673b6","Type":"ContainerStarted","Data":"5937d39267d3e8218ce39382aa4f2cd5bc7277ddb2f3e9d203fb60be7ae1d572"} Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.976113 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-646p8" event={"ID":"2e91d309-4c23-4d57-ad3f-8eb7634b9b04","Type":"ContainerStarted","Data":"2556098fc6e89af7c6c351ad10e0573f555efe8c133326fa21eefb11b9902401"} Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.978022 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.979233 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xlz27" event={"ID":"72e81d61-d517-412d-a5cb-100a5aac1ec3","Type":"ContainerStarted","Data":"c6c7dcc6a6a4840a78d37eb8074a4534402a4879ebd9e67c49333dcb254393dc"} Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.979263 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xlz27" event={"ID":"72e81d61-d517-412d-a5cb-100a5aac1ec3","Type":"ContainerStarted","Data":"5ae3302369c470ce8eac7086cd0906cc32a76eb7374908186697db3086d81c35"} Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.987450 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv4kz" event={"ID":"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4","Type":"ContainerStarted","Data":"443a8fa2069e99c030863c042183d939ecadb9561580d7eeef5f0284e38a1e60"} Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.987690 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv4kz" event={"ID":"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4","Type":"ContainerStarted","Data":"859c8c091e6f577156b2e09c7bc70e5f2effdba456b27f5fe65f338483749c71"} Sep 29 19:27:23 crc kubenswrapper[4741]: I0929 19:27:23.999135 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerStarted","Data":"bbe08b345bd3470436caafeeb50334a1f598e98ec9e674b5f2f5a87dbfd79645"} Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.005250 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" event={"ID":"00111435-439e-47b2-b8e8-08ee8fa40444","Type":"ContainerStarted","Data":"2d8f8feda7fc0e483eca7c54a56ebdda056d4d71041f5a6648457e62b461c04d"} Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.014091 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-xlz27" podStartSLOduration=2.014073498 podStartE2EDuration="2.014073498s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:23.997841234 +0000 UTC m=+1085.645630566" watchObservedRunningTime="2025-09-29 19:27:24.014073498 +0000 UTC m=+1085.661862830" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.017040 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gfpz8" event={"ID":"e172bee5-8543-4044-bf71-57e488d14fe4","Type":"ContainerStarted","Data":"4ed66f64565d1f1c11d4dc81367805d6d8aa89947eef157cd3134336fcb5aea7"} Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.021215 4741 generic.go:334] "Generic (PLEG): container finished" podID="6e632b91-9db1-4c5b-adb4-921af2dbcabe" containerID="dfaf7e3a34e0f4033a75467ed2df8a6c21f7363f1e5655eb34b5756ff06ff0fb" exitCode=0 Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.021257 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" event={"ID":"6e632b91-9db1-4c5b-adb4-921af2dbcabe","Type":"ContainerDied","Data":"dfaf7e3a34e0f4033a75467ed2df8a6c21f7363f1e5655eb34b5756ff06ff0fb"} Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.021290 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" event={"ID":"6e632b91-9db1-4c5b-adb4-921af2dbcabe","Type":"ContainerStarted","Data":"95396c7bae02e857156e3f66266d212f403f00f690f1a8b5b259bc015aab6b5d"} Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.028562 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rv4kz" podStartSLOduration=2.028543115 podStartE2EDuration="2.028543115s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:24.016913598 +0000 UTC m=+1085.664702930" watchObservedRunningTime="2025-09-29 19:27:24.028543115 +0000 UTC m=+1085.676332447" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.297074 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:24 crc kubenswrapper[4741]: W0929 19:27:24.314624 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88ace222_b0cb_492d_a0c8_3421e59a074e.slice/crio-d73ed9a86092e23c922a5ad98e47a86ef6d2b91a05e1b5e83f9ff923f0fae194 WatchSource:0}: Error finding container d73ed9a86092e23c922a5ad98e47a86ef6d2b91a05e1b5e83f9ff923f0fae194: Status 404 returned error can't find the container with id d73ed9a86092e23c922a5ad98e47a86ef6d2b91a05e1b5e83f9ff923f0fae194 Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.498415 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.664331 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-svc\") pod \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.664496 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-config\") pod \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.664523 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-sb\") pod \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.664556 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-nb\") pod \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.664589 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-swift-storage-0\") pod \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.664620 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkb4b\" (UniqueName: \"kubernetes.io/projected/6e632b91-9db1-4c5b-adb4-921af2dbcabe-kube-api-access-kkb4b\") pod \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\" (UID: \"6e632b91-9db1-4c5b-adb4-921af2dbcabe\") " Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.674109 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e632b91-9db1-4c5b-adb4-921af2dbcabe-kube-api-access-kkb4b" (OuterVolumeSpecName: "kube-api-access-kkb4b") pod "6e632b91-9db1-4c5b-adb4-921af2dbcabe" (UID: "6e632b91-9db1-4c5b-adb4-921af2dbcabe"). InnerVolumeSpecName "kube-api-access-kkb4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.703093 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6e632b91-9db1-4c5b-adb4-921af2dbcabe" (UID: "6e632b91-9db1-4c5b-adb4-921af2dbcabe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.706592 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-config" (OuterVolumeSpecName: "config") pod "6e632b91-9db1-4c5b-adb4-921af2dbcabe" (UID: "6e632b91-9db1-4c5b-adb4-921af2dbcabe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.728663 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e632b91-9db1-4c5b-adb4-921af2dbcabe" (UID: "6e632b91-9db1-4c5b-adb4-921af2dbcabe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.734460 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6e632b91-9db1-4c5b-adb4-921af2dbcabe" (UID: "6e632b91-9db1-4c5b-adb4-921af2dbcabe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.757131 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.766898 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkb4b\" (UniqueName: \"kubernetes.io/projected/6e632b91-9db1-4c5b-adb4-921af2dbcabe-kube-api-access-kkb4b\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.766927 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.766941 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.766950 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.766961 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.770069 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6e632b91-9db1-4c5b-adb4-921af2dbcabe" (UID: "6e632b91-9db1-4c5b-adb4-921af2dbcabe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:24 crc kubenswrapper[4741]: I0929 19:27:24.871886 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e632b91-9db1-4c5b-adb4-921af2dbcabe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.032971 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" event={"ID":"6e632b91-9db1-4c5b-adb4-921af2dbcabe","Type":"ContainerDied","Data":"95396c7bae02e857156e3f66266d212f403f00f690f1a8b5b259bc015aab6b5d"} Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.033020 4741 scope.go:117] "RemoveContainer" containerID="dfaf7e3a34e0f4033a75467ed2df8a6c21f7363f1e5655eb34b5756ff06ff0fb" Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.033068 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-xjq8s" Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.041123 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"88ace222-b0cb-492d-a0c8-3421e59a074e","Type":"ContainerStarted","Data":"d73ed9a86092e23c922a5ad98e47a86ef6d2b91a05e1b5e83f9ff923f0fae194"} Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.042358 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"85fad5ca-de2e-4c61-9dfc-ac890c21eca3","Type":"ContainerStarted","Data":"c1717401e068ccefd9f1637a8a3645c4963ec6f7f75c870bb435af809fcdb1f8"} Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.049550 4741 generic.go:334] "Generic (PLEG): container finished" podID="00111435-439e-47b2-b8e8-08ee8fa40444" containerID="8ccf185fed60f76c71590d2eb3b70147456c143c000e4ac7b9e5496f4ad963ab" exitCode=0 Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.050454 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" event={"ID":"00111435-439e-47b2-b8e8-08ee8fa40444","Type":"ContainerDied","Data":"8ccf185fed60f76c71590d2eb3b70147456c143c000e4ac7b9e5496f4ad963ab"} Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.216301 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-xjq8s"] Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.241499 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-xjq8s"] Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.750795 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.768530 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:27:25 crc kubenswrapper[4741]: I0929 19:27:25.820628 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:26 crc kubenswrapper[4741]: I0929 19:27:26.072996 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"85fad5ca-de2e-4c61-9dfc-ac890c21eca3","Type":"ContainerStarted","Data":"e06fdf54086e09f286a3519b96c9e0ab347af9e29aad3eb5a0ea37a6755958fc"} Sep 29 19:27:26 crc kubenswrapper[4741]: I0929 19:27:26.075275 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" event={"ID":"00111435-439e-47b2-b8e8-08ee8fa40444","Type":"ContainerStarted","Data":"5b46b99038b674100da22bc5da4251f6cd7fb3da90cde9218f7543a7494d97b6"} Sep 29 19:27:26 crc kubenswrapper[4741]: I0929 19:27:26.075407 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:26 crc kubenswrapper[4741]: I0929 19:27:26.081193 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"88ace222-b0cb-492d-a0c8-3421e59a074e","Type":"ContainerStarted","Data":"1e6e5fa5a8bb53852c4c02ca60fcf1eb84eb1eb1ae80ba895d2e541abfe7306a"} Sep 29 19:27:26 crc kubenswrapper[4741]: I0929 19:27:26.094782 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" podStartSLOduration=4.094763536 podStartE2EDuration="4.094763536s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:26.092365211 +0000 UTC m=+1087.740154543" watchObservedRunningTime="2025-09-29 19:27:26.094763536 +0000 UTC m=+1087.742552868" Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.096458 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e632b91-9db1-4c5b-adb4-921af2dbcabe" path="/var/lib/kubelet/pods/6e632b91-9db1-4c5b-adb4-921af2dbcabe/volumes" Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.096956 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-log" containerID="cri-o://1e6e5fa5a8bb53852c4c02ca60fcf1eb84eb1eb1ae80ba895d2e541abfe7306a" gracePeriod=30 Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.097352 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"88ace222-b0cb-492d-a0c8-3421e59a074e","Type":"ContainerStarted","Data":"6d878ce391b46bc3c88d1943376e6662e9ab603b28722673c6f8855d90433430"} Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.097384 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-httpd" containerID="cri-o://6d878ce391b46bc3c88d1943376e6662e9ab603b28722673c6f8855d90433430" gracePeriod=30 Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.100129 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"85fad5ca-de2e-4c61-9dfc-ac890c21eca3","Type":"ContainerStarted","Data":"b7dfd90ee63b3d185c04b5ac8eddb5907c099ff48df492d5a9705fd2a59300e7"} Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.100380 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-log" containerID="cri-o://e06fdf54086e09f286a3519b96c9e0ab347af9e29aad3eb5a0ea37a6755958fc" gracePeriod=30 Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.100469 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-httpd" containerID="cri-o://b7dfd90ee63b3d185c04b5ac8eddb5907c099ff48df492d5a9705fd2a59300e7" gracePeriod=30 Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.131936 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.131919065 podStartE2EDuration="5.131919065s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:27.117343014 +0000 UTC m=+1088.765132346" watchObservedRunningTime="2025-09-29 19:27:27.131919065 +0000 UTC m=+1088.779708397" Sep 29 19:27:27 crc kubenswrapper[4741]: I0929 19:27:27.152996 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.152978551 podStartE2EDuration="5.152978551s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:27.144498283 +0000 UTC m=+1088.792287615" watchObservedRunningTime="2025-09-29 19:27:27.152978551 +0000 UTC m=+1088.800767883" Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.109578 4741 generic.go:334] "Generic (PLEG): container finished" podID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerID="6d878ce391b46bc3c88d1943376e6662e9ab603b28722673c6f8855d90433430" exitCode=0 Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.110058 4741 generic.go:334] "Generic (PLEG): container finished" podID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerID="1e6e5fa5a8bb53852c4c02ca60fcf1eb84eb1eb1ae80ba895d2e541abfe7306a" exitCode=143 Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.109620 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"88ace222-b0cb-492d-a0c8-3421e59a074e","Type":"ContainerDied","Data":"6d878ce391b46bc3c88d1943376e6662e9ab603b28722673c6f8855d90433430"} Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.110143 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"88ace222-b0cb-492d-a0c8-3421e59a074e","Type":"ContainerDied","Data":"1e6e5fa5a8bb53852c4c02ca60fcf1eb84eb1eb1ae80ba895d2e541abfe7306a"} Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.112153 4741 generic.go:334] "Generic (PLEG): container finished" podID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerID="b7dfd90ee63b3d185c04b5ac8eddb5907c099ff48df492d5a9705fd2a59300e7" exitCode=0 Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.112190 4741 generic.go:334] "Generic (PLEG): container finished" podID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerID="e06fdf54086e09f286a3519b96c9e0ab347af9e29aad3eb5a0ea37a6755958fc" exitCode=143 Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.112202 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"85fad5ca-de2e-4c61-9dfc-ac890c21eca3","Type":"ContainerDied","Data":"b7dfd90ee63b3d185c04b5ac8eddb5907c099ff48df492d5a9705fd2a59300e7"} Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.112226 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"85fad5ca-de2e-4c61-9dfc-ac890c21eca3","Type":"ContainerDied","Data":"e06fdf54086e09f286a3519b96c9e0ab347af9e29aad3eb5a0ea37a6755958fc"} Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.116029 4741 generic.go:334] "Generic (PLEG): container finished" podID="c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" containerID="443a8fa2069e99c030863c042183d939ecadb9561580d7eeef5f0284e38a1e60" exitCode=0 Sep 29 19:27:28 crc kubenswrapper[4741]: I0929 19:27:28.116082 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv4kz" event={"ID":"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4","Type":"ContainerDied","Data":"443a8fa2069e99c030863c042183d939ecadb9561580d7eeef5f0284e38a1e60"} Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.913162 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.921680 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.938841 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992325 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lhqs\" (UniqueName: \"kubernetes.io/projected/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-kube-api-access-4lhqs\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992370 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-scripts\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992415 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-logs\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992467 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-httpd-run\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992500 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-scripts\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992519 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-config-data\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992540 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-combined-ca-bundle\") pod \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992558 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-combined-ca-bundle\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992584 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-logs\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992620 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992656 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-config-data\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992672 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-httpd-run\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992710 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-public-tls-certs\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992729 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-credential-keys\") pod \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992758 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpzz4\" (UniqueName: \"kubernetes.io/projected/88ace222-b0cb-492d-a0c8-3421e59a074e-kube-api-access-qpzz4\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992803 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-config-data\") pod \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992818 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs8v9\" (UniqueName: \"kubernetes.io/projected/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-kube-api-access-bs8v9\") pod \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992831 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-scripts\") pod \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992866 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-combined-ca-bundle\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992887 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"88ace222-b0cb-492d-a0c8-3421e59a074e\" (UID: \"88ace222-b0cb-492d-a0c8-3421e59a074e\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992903 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-fernet-keys\") pod \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\" (UID: \"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.992927 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-internal-tls-certs\") pod \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\" (UID: \"85fad5ca-de2e-4c61-9dfc-ac890c21eca3\") " Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.993016 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-logs" (OuterVolumeSpecName: "logs") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.993223 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:29 crc kubenswrapper[4741]: I0929 19:27:29.993791 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.000013 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-scripts" (OuterVolumeSpecName: "scripts") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.000504 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ace222-b0cb-492d-a0c8-3421e59a074e-kube-api-access-qpzz4" (OuterVolumeSpecName: "kube-api-access-qpzz4") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "kube-api-access-qpzz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.000764 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-scripts" (OuterVolumeSpecName: "scripts") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.001474 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.001615 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-scripts" (OuterVolumeSpecName: "scripts") pod "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" (UID: "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.001817 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.002478 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-logs" (OuterVolumeSpecName: "logs") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.003864 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-kube-api-access-bs8v9" (OuterVolumeSpecName: "kube-api-access-bs8v9") pod "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" (UID: "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4"). InnerVolumeSpecName "kube-api-access-bs8v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.011735 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.016592 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" (UID: "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.024179 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-kube-api-access-4lhqs" (OuterVolumeSpecName: "kube-api-access-4lhqs") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "kube-api-access-4lhqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.024620 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" (UID: "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.036640 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-config-data" (OuterVolumeSpecName: "config-data") pod "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" (UID: "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.038818 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.039353 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" (UID: "c484e98c-bf2d-4f4c-89fe-2ecd464e31a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.066359 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.079767 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-config-data" (OuterVolumeSpecName: "config-data") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.088627 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "88ace222-b0cb-492d-a0c8-3421e59a074e" (UID: "88ace222-b0cb-492d-a0c8-3421e59a074e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.092083 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097696 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097722 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097732 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097741 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097752 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097761 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097779 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097788 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88ace222-b0cb-492d-a0c8-3421e59a074e-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097797 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097805 4741 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097813 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpzz4\" (UniqueName: \"kubernetes.io/projected/88ace222-b0cb-492d-a0c8-3421e59a074e-kube-api-access-qpzz4\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097824 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs8v9\" (UniqueName: \"kubernetes.io/projected/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-kube-api-access-bs8v9\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097832 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097840 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097848 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ace222-b0cb-492d-a0c8-3421e59a074e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097860 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097869 4741 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097877 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097885 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lhqs\" (UniqueName: \"kubernetes.io/projected/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-kube-api-access-4lhqs\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.097892 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.131197 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.132219 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-config-data" (OuterVolumeSpecName: "config-data") pod "85fad5ca-de2e-4c61-9dfc-ac890c21eca3" (UID: "85fad5ca-de2e-4c61-9dfc-ac890c21eca3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.134924 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.162862 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"85fad5ca-de2e-4c61-9dfc-ac890c21eca3","Type":"ContainerDied","Data":"c1717401e068ccefd9f1637a8a3645c4963ec6f7f75c870bb435af809fcdb1f8"} Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.162913 4741 scope.go:117] "RemoveContainer" containerID="b7dfd90ee63b3d185c04b5ac8eddb5907c099ff48df492d5a9705fd2a59300e7" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.162974 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.170199 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv4kz" event={"ID":"c484e98c-bf2d-4f4c-89fe-2ecd464e31a4","Type":"ContainerDied","Data":"859c8c091e6f577156b2e09c7bc70e5f2effdba456b27f5fe65f338483749c71"} Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.170230 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="859c8c091e6f577156b2e09c7bc70e5f2effdba456b27f5fe65f338483749c71" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.170579 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv4kz" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.173641 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"88ace222-b0cb-492d-a0c8-3421e59a074e","Type":"ContainerDied","Data":"d73ed9a86092e23c922a5ad98e47a86ef6d2b91a05e1b5e83f9ff923f0fae194"} Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.173723 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.199432 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.199497 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.199514 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fad5ca-de2e-4c61-9dfc-ac890c21eca3-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.221453 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.232359 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.246135 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rv4kz"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.269352 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.281158 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rv4kz"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.290614 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.298044 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: E0929 19:27:30.298489 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e632b91-9db1-4c5b-adb4-921af2dbcabe" containerName="init" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.298508 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e632b91-9db1-4c5b-adb4-921af2dbcabe" containerName="init" Sep 29 19:27:30 crc kubenswrapper[4741]: E0929 19:27:30.298526 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-httpd" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.298533 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-httpd" Sep 29 19:27:30 crc kubenswrapper[4741]: E0929 19:27:30.298554 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" containerName="keystone-bootstrap" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.298562 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" containerName="keystone-bootstrap" Sep 29 19:27:30 crc kubenswrapper[4741]: E0929 19:27:30.298581 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-log" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.298590 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-log" Sep 29 19:27:30 crc kubenswrapper[4741]: E0929 19:27:30.298603 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-httpd" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.298611 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-httpd" Sep 29 19:27:30 crc kubenswrapper[4741]: E0929 19:27:30.298624 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-log" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.298632 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-log" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.299154 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" containerName="keystone-bootstrap" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.299177 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-httpd" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.299189 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e632b91-9db1-4c5b-adb4-921af2dbcabe" containerName="init" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.299201 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-log" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.299214 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" containerName="glance-httpd" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.299223 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" containerName="glance-log" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.301686 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.304745 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.304805 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-h9mn9" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.304763 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.304953 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.322613 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.324026 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.332652 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.332852 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.340521 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.346603 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.356168 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-s9tpc"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.357288 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.361360 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.361711 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kg6nl" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.361854 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.361976 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.368800 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s9tpc"] Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412164 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412278 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412329 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-config-data\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412352 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-credential-keys\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412403 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-config-data\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412429 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412484 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-fernet-keys\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412536 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bd6q\" (UniqueName: \"kubernetes.io/projected/9e4efdb3-fa44-4d1b-b62c-007252565734-kube-api-access-6bd6q\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412567 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-logs\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412587 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-scripts\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412624 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-scripts\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412653 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njxt6\" (UniqueName: \"kubernetes.io/projected/965471c2-5bf3-441c-a9de-1df141e5641e-kube-api-access-njxt6\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412697 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-logs\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412732 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412786 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412810 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj8dv\" (UniqueName: \"kubernetes.io/projected/d68f4aea-7345-4129-b25d-8d04773f9ebd-kube-api-access-qj8dv\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412870 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-combined-ca-bundle\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412897 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412938 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412960 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.412989 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.413042 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514171 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514223 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-config-data\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514243 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-config-data\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514260 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-credential-keys\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514279 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514308 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-fernet-keys\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514334 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bd6q\" (UniqueName: \"kubernetes.io/projected/9e4efdb3-fa44-4d1b-b62c-007252565734-kube-api-access-6bd6q\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514354 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-logs\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514370 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-scripts\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514385 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-scripts\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514421 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njxt6\" (UniqueName: \"kubernetes.io/projected/965471c2-5bf3-441c-a9de-1df141e5641e-kube-api-access-njxt6\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514436 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-logs\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514455 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514475 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514490 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj8dv\" (UniqueName: \"kubernetes.io/projected/d68f4aea-7345-4129-b25d-8d04773f9ebd-kube-api-access-qj8dv\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514509 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-combined-ca-bundle\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514528 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514542 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514556 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514575 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514599 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.514617 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.517626 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.518580 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-config-data\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.518750 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-logs\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.520008 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.520209 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.520692 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-logs\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.521364 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-fernet-keys\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.521610 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-config-data\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.521659 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.521936 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.522139 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.526634 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.529886 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.532430 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj8dv\" (UniqueName: \"kubernetes.io/projected/d68f4aea-7345-4129-b25d-8d04773f9ebd-kube-api-access-qj8dv\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.532525 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bd6q\" (UniqueName: \"kubernetes.io/projected/9e4efdb3-fa44-4d1b-b62c-007252565734-kube-api-access-6bd6q\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.534303 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-combined-ca-bundle\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.534681 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-scripts\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.535675 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-scripts\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.536410 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.537622 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.540926 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-credential-keys\") pod \"keystone-bootstrap-s9tpc\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.542017 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njxt6\" (UniqueName: \"kubernetes.io/projected/965471c2-5bf3-441c-a9de-1df141e5641e-kube-api-access-njxt6\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.555308 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.561050 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.641132 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.649772 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:30 crc kubenswrapper[4741]: I0929 19:27:30.679591 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:31 crc kubenswrapper[4741]: I0929 19:27:31.100734 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85fad5ca-de2e-4c61-9dfc-ac890c21eca3" path="/var/lib/kubelet/pods/85fad5ca-de2e-4c61-9dfc-ac890c21eca3/volumes" Sep 29 19:27:31 crc kubenswrapper[4741]: I0929 19:27:31.101809 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ace222-b0cb-492d-a0c8-3421e59a074e" path="/var/lib/kubelet/pods/88ace222-b0cb-492d-a0c8-3421e59a074e/volumes" Sep 29 19:27:31 crc kubenswrapper[4741]: I0929 19:27:31.102550 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c484e98c-bf2d-4f4c-89fe-2ecd464e31a4" path="/var/lib/kubelet/pods/c484e98c-bf2d-4f4c-89fe-2ecd464e31a4/volumes" Sep 29 19:27:33 crc kubenswrapper[4741]: I0929 19:27:33.258554 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:27:33 crc kubenswrapper[4741]: I0929 19:27:33.318632 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-77nb7"] Sep 29 19:27:33 crc kubenswrapper[4741]: I0929 19:27:33.318902 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="dnsmasq-dns" containerID="cri-o://8866fd403a8f411ea9d2d824752f67513dcb07bc65ebf9422999d845c9b65f48" gracePeriod=10 Sep 29 19:27:34 crc kubenswrapper[4741]: I0929 19:27:34.222863 4741 generic.go:334] "Generic (PLEG): container finished" podID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerID="8866fd403a8f411ea9d2d824752f67513dcb07bc65ebf9422999d845c9b65f48" exitCode=0 Sep 29 19:27:34 crc kubenswrapper[4741]: I0929 19:27:34.223126 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" event={"ID":"2c0af0d0-0060-4afb-800b-c63b70e69882","Type":"ContainerDied","Data":"8866fd403a8f411ea9d2d824752f67513dcb07bc65ebf9422999d845c9b65f48"} Sep 29 19:27:41 crc kubenswrapper[4741]: I0929 19:27:41.552128 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: i/o timeout" Sep 29 19:27:43 crc kubenswrapper[4741]: E0929 19:27:43.530348 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Sep 29 19:27:43 crc kubenswrapper[4741]: E0929 19:27:43.531091 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q5fg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-646p8_openstack(2e91d309-4c23-4d57-ad3f-8eb7634b9b04): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:27:43 crc kubenswrapper[4741]: E0929 19:27:43.532221 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-646p8" podUID="2e91d309-4c23-4d57-ad3f-8eb7634b9b04" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.608561 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.682874 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-svc\") pod \"2c0af0d0-0060-4afb-800b-c63b70e69882\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.682959 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-swift-storage-0\") pod \"2c0af0d0-0060-4afb-800b-c63b70e69882\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.683027 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-config\") pod \"2c0af0d0-0060-4afb-800b-c63b70e69882\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.683108 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-nb\") pod \"2c0af0d0-0060-4afb-800b-c63b70e69882\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.683140 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-sb\") pod \"2c0af0d0-0060-4afb-800b-c63b70e69882\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.683166 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg2sr\" (UniqueName: \"kubernetes.io/projected/2c0af0d0-0060-4afb-800b-c63b70e69882-kube-api-access-vg2sr\") pod \"2c0af0d0-0060-4afb-800b-c63b70e69882\" (UID: \"2c0af0d0-0060-4afb-800b-c63b70e69882\") " Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.690568 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c0af0d0-0060-4afb-800b-c63b70e69882-kube-api-access-vg2sr" (OuterVolumeSpecName: "kube-api-access-vg2sr") pod "2c0af0d0-0060-4afb-800b-c63b70e69882" (UID: "2c0af0d0-0060-4afb-800b-c63b70e69882"). InnerVolumeSpecName "kube-api-access-vg2sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.724491 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2c0af0d0-0060-4afb-800b-c63b70e69882" (UID: "2c0af0d0-0060-4afb-800b-c63b70e69882"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.727813 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-config" (OuterVolumeSpecName: "config") pod "2c0af0d0-0060-4afb-800b-c63b70e69882" (UID: "2c0af0d0-0060-4afb-800b-c63b70e69882"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.727845 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2c0af0d0-0060-4afb-800b-c63b70e69882" (UID: "2c0af0d0-0060-4afb-800b-c63b70e69882"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.731071 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2c0af0d0-0060-4afb-800b-c63b70e69882" (UID: "2c0af0d0-0060-4afb-800b-c63b70e69882"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.742414 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2c0af0d0-0060-4afb-800b-c63b70e69882" (UID: "2c0af0d0-0060-4afb-800b-c63b70e69882"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.784520 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.784551 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.784562 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.784570 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.784578 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c0af0d0-0060-4afb-800b-c63b70e69882-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:43 crc kubenswrapper[4741]: I0929 19:27:43.784587 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg2sr\" (UniqueName: \"kubernetes.io/projected/2c0af0d0-0060-4afb-800b-c63b70e69882-kube-api-access-vg2sr\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:43 crc kubenswrapper[4741]: E0929 19:27:43.881424 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Sep 29 19:27:43 crc kubenswrapper[4741]: E0929 19:27:43.881593 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n587h5bdh656h76h656hd7hcbh5c4h54fh56fh9bhbchc7h96h54bh55ch8h64bhf5h5cbh58fh64h5cfh564h5fbh5d6h685hd7h77h54fh65hb5q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zxjtt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(34dc5bf3-42dd-47ec-ba42-7422df4dde51): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:27:44 crc kubenswrapper[4741]: I0929 19:27:44.298623 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" event={"ID":"2c0af0d0-0060-4afb-800b-c63b70e69882","Type":"ContainerDied","Data":"c8b093fff128673bcf6e8f30072cbe2ecbcfb4058a0995b2315edf23cf313983"} Sep 29 19:27:44 crc kubenswrapper[4741]: I0929 19:27:44.298927 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" Sep 29 19:27:44 crc kubenswrapper[4741]: I0929 19:27:44.300756 4741 generic.go:334] "Generic (PLEG): container finished" podID="72e81d61-d517-412d-a5cb-100a5aac1ec3" containerID="c6c7dcc6a6a4840a78d37eb8074a4534402a4879ebd9e67c49333dcb254393dc" exitCode=0 Sep 29 19:27:44 crc kubenswrapper[4741]: I0929 19:27:44.301427 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xlz27" event={"ID":"72e81d61-d517-412d-a5cb-100a5aac1ec3","Type":"ContainerDied","Data":"c6c7dcc6a6a4840a78d37eb8074a4534402a4879ebd9e67c49333dcb254393dc"} Sep 29 19:27:44 crc kubenswrapper[4741]: E0929 19:27:44.303147 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-646p8" podUID="2e91d309-4c23-4d57-ad3f-8eb7634b9b04" Sep 29 19:27:44 crc kubenswrapper[4741]: I0929 19:27:44.348593 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-77nb7"] Sep 29 19:27:44 crc kubenswrapper[4741]: I0929 19:27:44.355317 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-77nb7"] Sep 29 19:27:44 crc kubenswrapper[4741]: I0929 19:27:44.889941 4741 scope.go:117] "RemoveContainer" containerID="e06fdf54086e09f286a3519b96c9e0ab347af9e29aad3eb5a0ea37a6755958fc" Sep 29 19:27:44 crc kubenswrapper[4741]: E0929 19:27:44.911120 4741 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Sep 29 19:27:44 crc kubenswrapper[4741]: E0929 19:27:44.911518 4741 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-67s8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-gfpz8_openstack(e172bee5-8543-4044-bf71-57e488d14fe4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 19:27:44 crc kubenswrapper[4741]: E0929 19:27:44.912721 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-gfpz8" podUID="e172bee5-8543-4044-bf71-57e488d14fe4" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.055170 4741 scope.go:117] "RemoveContainer" containerID="6d878ce391b46bc3c88d1943376e6662e9ab603b28722673c6f8855d90433430" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.087113 4741 scope.go:117] "RemoveContainer" containerID="1e6e5fa5a8bb53852c4c02ca60fcf1eb84eb1eb1ae80ba895d2e541abfe7306a" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.100260 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" path="/var/lib/kubelet/pods/2c0af0d0-0060-4afb-800b-c63b70e69882/volumes" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.122359 4741 scope.go:117] "RemoveContainer" containerID="8866fd403a8f411ea9d2d824752f67513dcb07bc65ebf9422999d845c9b65f48" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.138901 4741 scope.go:117] "RemoveContainer" containerID="ab5fc5912468e5b08ef1f395eb7eb27bc44632e5a7ac4a11539ac0f08f30db74" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.310850 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hjb7q" event={"ID":"13ca2f00-5599-4912-83dd-afbe369673b6","Type":"ContainerStarted","Data":"bd272abdec76fda4950b811de1c1e048d2e279faf0d1fac0874af53e2d4b1e91"} Sep 29 19:27:45 crc kubenswrapper[4741]: E0929 19:27:45.317061 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-gfpz8" podUID="e172bee5-8543-4044-bf71-57e488d14fe4" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.352145 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-hjb7q" podStartSLOduration=3.373689563 podStartE2EDuration="23.352127874s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="2025-09-29 19:27:23.908604172 +0000 UTC m=+1085.556393504" lastFinishedPulling="2025-09-29 19:27:43.887042473 +0000 UTC m=+1105.534831815" observedRunningTime="2025-09-29 19:27:45.335095246 +0000 UTC m=+1106.982884588" watchObservedRunningTime="2025-09-29 19:27:45.352127874 +0000 UTC m=+1106.999917206" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.442729 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s9tpc"] Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.460593 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:27:45 crc kubenswrapper[4741]: W0929 19:27:45.648018 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e4efdb3_fa44_4d1b_b62c_007252565734.slice/crio-b87a3e6d8c2f8a2df169a6be07306c13df53bf202806a30f1bb20810980ba423 WatchSource:0}: Error finding container b87a3e6d8c2f8a2df169a6be07306c13df53bf202806a30f1bb20810980ba423: Status 404 returned error can't find the container with id b87a3e6d8c2f8a2df169a6be07306c13df53bf202806a30f1bb20810980ba423 Sep 29 19:27:45 crc kubenswrapper[4741]: W0929 19:27:45.661839 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod965471c2_5bf3_441c_a9de_1df141e5641e.slice/crio-4d7b1baeb9848bc71d0c5143c2381678f682f4749160024d689041b63b94411e WatchSource:0}: Error finding container 4d7b1baeb9848bc71d0c5143c2381678f682f4749160024d689041b63b94411e: Status 404 returned error can't find the container with id 4d7b1baeb9848bc71d0c5143c2381678f682f4749160024d689041b63b94411e Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.846133 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.920492 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-config\") pod \"72e81d61-d517-412d-a5cb-100a5aac1ec3\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.920572 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-combined-ca-bundle\") pod \"72e81d61-d517-412d-a5cb-100a5aac1ec3\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.921511 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc4nn\" (UniqueName: \"kubernetes.io/projected/72e81d61-d517-412d-a5cb-100a5aac1ec3-kube-api-access-sc4nn\") pod \"72e81d61-d517-412d-a5cb-100a5aac1ec3\" (UID: \"72e81d61-d517-412d-a5cb-100a5aac1ec3\") " Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.927928 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e81d61-d517-412d-a5cb-100a5aac1ec3-kube-api-access-sc4nn" (OuterVolumeSpecName: "kube-api-access-sc4nn") pod "72e81d61-d517-412d-a5cb-100a5aac1ec3" (UID: "72e81d61-d517-412d-a5cb-100a5aac1ec3"). InnerVolumeSpecName "kube-api-access-sc4nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.945205 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-config" (OuterVolumeSpecName: "config") pod "72e81d61-d517-412d-a5cb-100a5aac1ec3" (UID: "72e81d61-d517-412d-a5cb-100a5aac1ec3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:45 crc kubenswrapper[4741]: I0929 19:27:45.955730 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72e81d61-d517-412d-a5cb-100a5aac1ec3" (UID: "72e81d61-d517-412d-a5cb-100a5aac1ec3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.026373 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc4nn\" (UniqueName: \"kubernetes.io/projected/72e81d61-d517-412d-a5cb-100a5aac1ec3-kube-api-access-sc4nn\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.026449 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.026470 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e81d61-d517-412d-a5cb-100a5aac1ec3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.339598 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerStarted","Data":"5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d"} Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.341521 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s9tpc" event={"ID":"9e4efdb3-fa44-4d1b-b62c-007252565734","Type":"ContainerStarted","Data":"f0f5d8ecf900696782222753fa68359e26bb6b0234dc48e8dad493ca3f5967bc"} Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.341570 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s9tpc" event={"ID":"9e4efdb3-fa44-4d1b-b62c-007252565734","Type":"ContainerStarted","Data":"b87a3e6d8c2f8a2df169a6be07306c13df53bf202806a30f1bb20810980ba423"} Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.345028 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965471c2-5bf3-441c-a9de-1df141e5641e","Type":"ContainerStarted","Data":"cb9ae15c067183744cd6dc00963bb2688701cfa6bf3cfda30d99cddb222a8a6e"} Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.345063 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965471c2-5bf3-441c-a9de-1df141e5641e","Type":"ContainerStarted","Data":"4d7b1baeb9848bc71d0c5143c2381678f682f4749160024d689041b63b94411e"} Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.351488 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xlz27" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.351515 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xlz27" event={"ID":"72e81d61-d517-412d-a5cb-100a5aac1ec3","Type":"ContainerDied","Data":"5ae3302369c470ce8eac7086cd0906cc32a76eb7374908186697db3086d81c35"} Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.351552 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ae3302369c470ce8eac7086cd0906cc32a76eb7374908186697db3086d81c35" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.378585 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-s9tpc" podStartSLOduration=16.378569434 podStartE2EDuration="16.378569434s" podCreationTimestamp="2025-09-29 19:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:46.364165838 +0000 UTC m=+1108.011955170" watchObservedRunningTime="2025-09-29 19:27:46.378569434 +0000 UTC m=+1108.026358766" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.455935 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:27:46 crc kubenswrapper[4741]: W0929 19:27:46.461710 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd68f4aea_7345_4129_b25d_8d04773f9ebd.slice/crio-763617a7a4685e1563da99332b0ae71a7874bd46cabfe82b8a58f983d346fea7 WatchSource:0}: Error finding container 763617a7a4685e1563da99332b0ae71a7874bd46cabfe82b8a58f983d346fea7: Status 404 returned error can't find the container with id 763617a7a4685e1563da99332b0ae71a7874bd46cabfe82b8a58f983d346fea7 Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.557426 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-77nb7" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: i/o timeout" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.565794 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wtx87"] Sep 29 19:27:46 crc kubenswrapper[4741]: E0929 19:27:46.566153 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e81d61-d517-412d-a5cb-100a5aac1ec3" containerName="neutron-db-sync" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.566163 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e81d61-d517-412d-a5cb-100a5aac1ec3" containerName="neutron-db-sync" Sep 29 19:27:46 crc kubenswrapper[4741]: E0929 19:27:46.566189 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="init" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.566194 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="init" Sep 29 19:27:46 crc kubenswrapper[4741]: E0929 19:27:46.566207 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="dnsmasq-dns" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.566213 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="dnsmasq-dns" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.566357 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e81d61-d517-412d-a5cb-100a5aac1ec3" containerName="neutron-db-sync" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.566371 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c0af0d0-0060-4afb-800b-c63b70e69882" containerName="dnsmasq-dns" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.567181 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.608941 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wtx87"] Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.639431 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-svc\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.640922 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.640983 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.641048 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-config\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.641134 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q52mq\" (UniqueName: \"kubernetes.io/projected/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-kube-api-access-q52mq\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.641279 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.728657 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-86bbf6b688-lcffw"] Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.737273 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.740989 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.741199 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86bbf6b688-lcffw"] Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.741524 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-c9t62" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.741742 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.742261 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-svc\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.742297 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.742328 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.742365 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-config\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.742423 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q52mq\" (UniqueName: \"kubernetes.io/projected/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-kube-api-access-q52mq\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.742463 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.743241 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.743933 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-svc\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.744133 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.744454 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-config\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.744721 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.746803 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.779269 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q52mq\" (UniqueName: \"kubernetes.io/projected/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-kube-api-access-q52mq\") pod \"dnsmasq-dns-55f844cf75-wtx87\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.845217 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-combined-ca-bundle\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.845326 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-config\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.845360 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-ovndb-tls-certs\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.845378 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnqtv\" (UniqueName: \"kubernetes.io/projected/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-kube-api-access-tnqtv\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.845444 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-httpd-config\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.896206 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.946604 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-config\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.946657 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-ovndb-tls-certs\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.946681 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnqtv\" (UniqueName: \"kubernetes.io/projected/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-kube-api-access-tnqtv\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.946711 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-httpd-config\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.946753 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-combined-ca-bundle\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.953739 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-config\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.957147 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-combined-ca-bundle\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.964171 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnqtv\" (UniqueName: \"kubernetes.io/projected/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-kube-api-access-tnqtv\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.967151 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-ovndb-tls-certs\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:46 crc kubenswrapper[4741]: I0929 19:27:46.986002 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-httpd-config\") pod \"neutron-86bbf6b688-lcffw\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.071082 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.368819 4741 generic.go:334] "Generic (PLEG): container finished" podID="13ca2f00-5599-4912-83dd-afbe369673b6" containerID="bd272abdec76fda4950b811de1c1e048d2e279faf0d1fac0874af53e2d4b1e91" exitCode=0 Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.369281 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hjb7q" event={"ID":"13ca2f00-5599-4912-83dd-afbe369673b6","Type":"ContainerDied","Data":"bd272abdec76fda4950b811de1c1e048d2e279faf0d1fac0874af53e2d4b1e91"} Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.385669 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965471c2-5bf3-441c-a9de-1df141e5641e","Type":"ContainerStarted","Data":"49827be4e3437ca826ee30ca5dafe5dba420363ceccdaa1f5c0f8ad732401de7"} Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.388418 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d68f4aea-7345-4129-b25d-8d04773f9ebd","Type":"ContainerStarted","Data":"3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87"} Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.388458 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d68f4aea-7345-4129-b25d-8d04773f9ebd","Type":"ContainerStarted","Data":"763617a7a4685e1563da99332b0ae71a7874bd46cabfe82b8a58f983d346fea7"} Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.452251 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wtx87"] Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.466890 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.466871489 podStartE2EDuration="17.466871489s" podCreationTimestamp="2025-09-29 19:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:47.456651067 +0000 UTC m=+1109.104440399" watchObservedRunningTime="2025-09-29 19:27:47.466871489 +0000 UTC m=+1109.114660821" Sep 29 19:27:47 crc kubenswrapper[4741]: I0929 19:27:47.674769 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86bbf6b688-lcffw"] Sep 29 19:27:47 crc kubenswrapper[4741]: W0929 19:27:47.692969 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddccc1fd0_0b9f_41a0_89fe_8dcae0f281c1.slice/crio-b7426772344e481587347259225cdde7277e7a3d2a9b97397bc8febfb8493777 WatchSource:0}: Error finding container b7426772344e481587347259225cdde7277e7a3d2a9b97397bc8febfb8493777: Status 404 returned error can't find the container with id b7426772344e481587347259225cdde7277e7a3d2a9b97397bc8febfb8493777 Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.404511 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d68f4aea-7345-4129-b25d-8d04773f9ebd","Type":"ContainerStarted","Data":"38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0"} Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.437767 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=18.437748123 podStartE2EDuration="18.437748123s" podCreationTimestamp="2025-09-29 19:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:48.433975793 +0000 UTC m=+1110.081765135" watchObservedRunningTime="2025-09-29 19:27:48.437748123 +0000 UTC m=+1110.085537455" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.444085 4741 generic.go:334] "Generic (PLEG): container finished" podID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerID="d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200" exitCode=0 Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.444150 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" event={"ID":"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32","Type":"ContainerDied","Data":"d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200"} Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.444176 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" event={"ID":"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32","Type":"ContainerStarted","Data":"9c1ea77995031fcbca61b25ce56741ac0141e1383b317e43c946672661dba0c6"} Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.447982 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86bbf6b688-lcffw" event={"ID":"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1","Type":"ContainerStarted","Data":"e0350f3b0dc57e12e5f350135bdb6938f5c4bd6180fecfcf91eb04371e1d439b"} Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.448021 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86bbf6b688-lcffw" event={"ID":"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1","Type":"ContainerStarted","Data":"20400580a7b1bb2a730ab3cf2404a7dde79dceaf6d945e1e6553f0c844c4aa95"} Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.448034 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86bbf6b688-lcffw" event={"ID":"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1","Type":"ContainerStarted","Data":"b7426772344e481587347259225cdde7277e7a3d2a9b97397bc8febfb8493777"} Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.448853 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.677263 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-86bbf6b688-lcffw" podStartSLOduration=2.677239786 podStartE2EDuration="2.677239786s" podCreationTimestamp="2025-09-29 19:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:48.55404246 +0000 UTC m=+1110.201831812" watchObservedRunningTime="2025-09-29 19:27:48.677239786 +0000 UTC m=+1110.325029128" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.696461 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5dd77b8d5c-hw4v7"] Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.698335 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.701086 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.703034 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.710569 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dd77b8d5c-hw4v7"] Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.781482 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76fvb\" (UniqueName: \"kubernetes.io/projected/760672f8-7a2f-4ddb-8ebd-3664670a4adc-kube-api-access-76fvb\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.781586 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-combined-ca-bundle\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.781626 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-internal-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.781716 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-ovndb-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.781747 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.782244 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.782314 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-public-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.845490 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.883774 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-config-data\") pod \"13ca2f00-5599-4912-83dd-afbe369673b6\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.883846 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13ca2f00-5599-4912-83dd-afbe369673b6-logs\") pod \"13ca2f00-5599-4912-83dd-afbe369673b6\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.883912 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fss9\" (UniqueName: \"kubernetes.io/projected/13ca2f00-5599-4912-83dd-afbe369673b6-kube-api-access-5fss9\") pod \"13ca2f00-5599-4912-83dd-afbe369673b6\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.883977 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-combined-ca-bundle\") pod \"13ca2f00-5599-4912-83dd-afbe369673b6\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884057 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-scripts\") pod \"13ca2f00-5599-4912-83dd-afbe369673b6\" (UID: \"13ca2f00-5599-4912-83dd-afbe369673b6\") " Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884336 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-ovndb-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884365 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884459 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884478 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-public-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884508 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76fvb\" (UniqueName: \"kubernetes.io/projected/760672f8-7a2f-4ddb-8ebd-3664670a4adc-kube-api-access-76fvb\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884598 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-combined-ca-bundle\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.884626 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-internal-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.885471 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13ca2f00-5599-4912-83dd-afbe369673b6-logs" (OuterVolumeSpecName: "logs") pod "13ca2f00-5599-4912-83dd-afbe369673b6" (UID: "13ca2f00-5599-4912-83dd-afbe369673b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.893744 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-combined-ca-bundle\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.893796 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.894311 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-public-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.895456 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-ovndb-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.895769 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13ca2f00-5599-4912-83dd-afbe369673b6-kube-api-access-5fss9" (OuterVolumeSpecName: "kube-api-access-5fss9") pod "13ca2f00-5599-4912-83dd-afbe369673b6" (UID: "13ca2f00-5599-4912-83dd-afbe369673b6"). InnerVolumeSpecName "kube-api-access-5fss9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.899005 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-internal-tls-certs\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.899109 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-scripts" (OuterVolumeSpecName: "scripts") pod "13ca2f00-5599-4912-83dd-afbe369673b6" (UID: "13ca2f00-5599-4912-83dd-afbe369673b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.899408 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.917437 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76fvb\" (UniqueName: \"kubernetes.io/projected/760672f8-7a2f-4ddb-8ebd-3664670a4adc-kube-api-access-76fvb\") pod \"neutron-5dd77b8d5c-hw4v7\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.918409 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-config-data" (OuterVolumeSpecName: "config-data") pod "13ca2f00-5599-4912-83dd-afbe369673b6" (UID: "13ca2f00-5599-4912-83dd-afbe369673b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.940440 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13ca2f00-5599-4912-83dd-afbe369673b6" (UID: "13ca2f00-5599-4912-83dd-afbe369673b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.986624 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.987057 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13ca2f00-5599-4912-83dd-afbe369673b6-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.987070 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fss9\" (UniqueName: \"kubernetes.io/projected/13ca2f00-5599-4912-83dd-afbe369673b6-kube-api-access-5fss9\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.987206 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:48 crc kubenswrapper[4741]: I0929 19:27:48.987216 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ca2f00-5599-4912-83dd-afbe369673b6-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.141850 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.459677 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hjb7q" event={"ID":"13ca2f00-5599-4912-83dd-afbe369673b6","Type":"ContainerDied","Data":"5937d39267d3e8218ce39382aa4f2cd5bc7277ddb2f3e9d203fb60be7ae1d572"} Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.459713 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5937d39267d3e8218ce39382aa4f2cd5bc7277ddb2f3e9d203fb60be7ae1d572" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.459722 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hjb7q" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.461996 4741 generic.go:334] "Generic (PLEG): container finished" podID="9e4efdb3-fa44-4d1b-b62c-007252565734" containerID="f0f5d8ecf900696782222753fa68359e26bb6b0234dc48e8dad493ca3f5967bc" exitCode=0 Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.462054 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s9tpc" event={"ID":"9e4efdb3-fa44-4d1b-b62c-007252565734","Type":"ContainerDied","Data":"f0f5d8ecf900696782222753fa68359e26bb6b0234dc48e8dad493ca3f5967bc"} Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.485619 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" event={"ID":"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32","Type":"ContainerStarted","Data":"2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4"} Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.485669 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.500882 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c686d89f8-xwpjv"] Sep 29 19:27:49 crc kubenswrapper[4741]: E0929 19:27:49.501250 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ca2f00-5599-4912-83dd-afbe369673b6" containerName="placement-db-sync" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.501268 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ca2f00-5599-4912-83dd-afbe369673b6" containerName="placement-db-sync" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.501468 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ca2f00-5599-4912-83dd-afbe369673b6" containerName="placement-db-sync" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.502365 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.504744 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.505180 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.505343 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.505534 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pwnw4" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.511521 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.537883 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c686d89f8-xwpjv"] Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.544195 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" podStartSLOduration=3.544003466 podStartE2EDuration="3.544003466s" podCreationTimestamp="2025-09-29 19:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:49.503756864 +0000 UTC m=+1111.151546186" watchObservedRunningTime="2025-09-29 19:27:49.544003466 +0000 UTC m=+1111.191792798" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.605012 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-internal-tls-certs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.605058 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-public-tls-certs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.605432 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-scripts\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.605510 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-config-data\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.605630 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-combined-ca-bundle\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.605783 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skkr5\" (UniqueName: \"kubernetes.io/projected/2be2676f-d882-424e-8a49-d8bffc23773e-kube-api-access-skkr5\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.606079 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2be2676f-d882-424e-8a49-d8bffc23773e-logs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.707863 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2be2676f-d882-424e-8a49-d8bffc23773e-logs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.707984 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-internal-tls-certs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.708031 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-public-tls-certs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.708137 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-scripts\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.708136 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2be2676f-d882-424e-8a49-d8bffc23773e-logs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.708170 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-config-data\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.708219 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-combined-ca-bundle\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.708281 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skkr5\" (UniqueName: \"kubernetes.io/projected/2be2676f-d882-424e-8a49-d8bffc23773e-kube-api-access-skkr5\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.713648 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-internal-tls-certs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.715914 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-public-tls-certs\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.715957 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-scripts\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.716190 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-combined-ca-bundle\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.718350 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-config-data\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.722878 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skkr5\" (UniqueName: \"kubernetes.io/projected/2be2676f-d882-424e-8a49-d8bffc23773e-kube-api-access-skkr5\") pod \"placement-c686d89f8-xwpjv\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:49 crc kubenswrapper[4741]: I0929 19:27:49.830751 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.642282 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.642733 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.650270 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.650321 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.686806 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.688310 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.751620 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:50 crc kubenswrapper[4741]: I0929 19:27:50.823053 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:51 crc kubenswrapper[4741]: I0929 19:27:51.503030 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:51 crc kubenswrapper[4741]: I0929 19:27:51.503357 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 19:27:51 crc kubenswrapper[4741]: I0929 19:27:51.503377 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 19:27:51 crc kubenswrapper[4741]: I0929 19:27:51.503403 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.335867 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.468294 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bd6q\" (UniqueName: \"kubernetes.io/projected/9e4efdb3-fa44-4d1b-b62c-007252565734-kube-api-access-6bd6q\") pod \"9e4efdb3-fa44-4d1b-b62c-007252565734\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.468662 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-scripts\") pod \"9e4efdb3-fa44-4d1b-b62c-007252565734\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.468739 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-credential-keys\") pod \"9e4efdb3-fa44-4d1b-b62c-007252565734\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.468818 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-fernet-keys\") pod \"9e4efdb3-fa44-4d1b-b62c-007252565734\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.468917 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-combined-ca-bundle\") pod \"9e4efdb3-fa44-4d1b-b62c-007252565734\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.469014 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-config-data\") pod \"9e4efdb3-fa44-4d1b-b62c-007252565734\" (UID: \"9e4efdb3-fa44-4d1b-b62c-007252565734\") " Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.475057 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-scripts" (OuterVolumeSpecName: "scripts") pod "9e4efdb3-fa44-4d1b-b62c-007252565734" (UID: "9e4efdb3-fa44-4d1b-b62c-007252565734"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.475479 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9e4efdb3-fa44-4d1b-b62c-007252565734" (UID: "9e4efdb3-fa44-4d1b-b62c-007252565734"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.475712 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e4efdb3-fa44-4d1b-b62c-007252565734-kube-api-access-6bd6q" (OuterVolumeSpecName: "kube-api-access-6bd6q") pod "9e4efdb3-fa44-4d1b-b62c-007252565734" (UID: "9e4efdb3-fa44-4d1b-b62c-007252565734"). InnerVolumeSpecName "kube-api-access-6bd6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.476487 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9e4efdb3-fa44-4d1b-b62c-007252565734" (UID: "9e4efdb3-fa44-4d1b-b62c-007252565734"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.499760 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e4efdb3-fa44-4d1b-b62c-007252565734" (UID: "9e4efdb3-fa44-4d1b-b62c-007252565734"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.516977 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-config-data" (OuterVolumeSpecName: "config-data") pod "9e4efdb3-fa44-4d1b-b62c-007252565734" (UID: "9e4efdb3-fa44-4d1b-b62c-007252565734"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.517425 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerStarted","Data":"7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437"} Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.519494 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s9tpc" event={"ID":"9e4efdb3-fa44-4d1b-b62c-007252565734","Type":"ContainerDied","Data":"b87a3e6d8c2f8a2df169a6be07306c13df53bf202806a30f1bb20810980ba423"} Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.519534 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b87a3e6d8c2f8a2df169a6be07306c13df53bf202806a30f1bb20810980ba423" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.519643 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s9tpc" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.570614 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.570642 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bd6q\" (UniqueName: \"kubernetes.io/projected/9e4efdb3-fa44-4d1b-b62c-007252565734-kube-api-access-6bd6q\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.570653 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.570661 4741 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.570669 4741 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.570677 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4efdb3-fa44-4d1b-b62c-007252565734-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.643688 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c686d89f8-xwpjv"] Sep 29 19:27:52 crc kubenswrapper[4741]: I0929 19:27:52.786718 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dd77b8d5c-hw4v7"] Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.416086 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-677b8f6f88-c7kf7"] Sep 29 19:27:53 crc kubenswrapper[4741]: E0929 19:27:53.416947 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4efdb3-fa44-4d1b-b62c-007252565734" containerName="keystone-bootstrap" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.416961 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4efdb3-fa44-4d1b-b62c-007252565734" containerName="keystone-bootstrap" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.417139 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4efdb3-fa44-4d1b-b62c-007252565734" containerName="keystone-bootstrap" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.417669 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.419854 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.420070 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.420242 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.420420 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kg6nl" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.420570 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.420683 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.438756 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-677b8f6f88-c7kf7"] Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.489723 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.489786 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.489878 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-combined-ca-bundle\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.489917 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.489945 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-public-tls-certs\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.490000 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.490053 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hklql\" (UniqueName: \"kubernetes.io/projected/c81db200-8fff-4a29-952c-4f8609dc8ac8-kube-api-access-hklql\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.490161 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-internal-tls-certs\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.531689 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c686d89f8-xwpjv" event={"ID":"2be2676f-d882-424e-8a49-d8bffc23773e","Type":"ContainerStarted","Data":"88822bc2858601e3da0fba4f6a6ca50dcd1d5c1fe288573a193d9458b638d155"} Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.531730 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c686d89f8-xwpjv" event={"ID":"2be2676f-d882-424e-8a49-d8bffc23773e","Type":"ContainerStarted","Data":"c7249b3b7837fc3520fcc2d531958323f2ab743d34b60e50c851a65328623db0"} Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.531740 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c686d89f8-xwpjv" event={"ID":"2be2676f-d882-424e-8a49-d8bffc23773e","Type":"ContainerStarted","Data":"fe9ff92122e26d5f80efee609332fb9b0ee005e5abfc50bf9d3618a7103d1922"} Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.531784 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.531804 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.536228 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd77b8d5c-hw4v7" event={"ID":"760672f8-7a2f-4ddb-8ebd-3664670a4adc","Type":"ContainerStarted","Data":"1700c1e978f1aa2ccdb75b0383b4cb4e5f89351b953d72fcb8ea36e16ead7085"} Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.536270 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd77b8d5c-hw4v7" event={"ID":"760672f8-7a2f-4ddb-8ebd-3664670a4adc","Type":"ContainerStarted","Data":"3d283ad6e8c834fde894dcf133b306ef1602160587b86f10edad93244d544e8f"} Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.536286 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd77b8d5c-hw4v7" event={"ID":"760672f8-7a2f-4ddb-8ebd-3664670a4adc","Type":"ContainerStarted","Data":"b86fa10f7eec29753146890f1e29da4af0d830a57cc9f2a854021a7e52abd618"} Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.537081 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.558222 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-c686d89f8-xwpjv" podStartSLOduration=4.55820591 podStartE2EDuration="4.55820591s" podCreationTimestamp="2025-09-29 19:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:53.556752864 +0000 UTC m=+1115.204542196" watchObservedRunningTime="2025-09-29 19:27:53.55820591 +0000 UTC m=+1115.205995242" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.587366 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5dd77b8d5c-hw4v7" podStartSLOduration=5.587345681 podStartE2EDuration="5.587345681s" podCreationTimestamp="2025-09-29 19:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:53.579949657 +0000 UTC m=+1115.227738989" watchObservedRunningTime="2025-09-29 19:27:53.587345681 +0000 UTC m=+1115.235135013" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.591944 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-combined-ca-bundle\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.592039 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.592091 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-public-tls-certs\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.592156 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.592260 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hklql\" (UniqueName: \"kubernetes.io/projected/c81db200-8fff-4a29-952c-4f8609dc8ac8-kube-api-access-hklql\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.592303 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-internal-tls-certs\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.592421 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.592458 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.596076 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.598517 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.601860 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-internal-tls-certs\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.619711 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hklql\" (UniqueName: \"kubernetes.io/projected/c81db200-8fff-4a29-952c-4f8609dc8ac8-kube-api-access-hklql\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.620359 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.622887 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.623482 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-combined-ca-bundle\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.635363 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-public-tls-certs\") pod \"keystone-677b8f6f88-c7kf7\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.739651 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:53 crc kubenswrapper[4741]: I0929 19:27:53.776670 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:54 crc kubenswrapper[4741]: I0929 19:27:54.224328 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-677b8f6f88-c7kf7"] Sep 29 19:27:54 crc kubenswrapper[4741]: I0929 19:27:54.569835 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677b8f6f88-c7kf7" event={"ID":"c81db200-8fff-4a29-952c-4f8609dc8ac8","Type":"ContainerStarted","Data":"18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34"} Sep 29 19:27:54 crc kubenswrapper[4741]: I0929 19:27:54.570168 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677b8f6f88-c7kf7" event={"ID":"c81db200-8fff-4a29-952c-4f8609dc8ac8","Type":"ContainerStarted","Data":"c64598fbf2851b5a62163b3e44ad7be97327792b5cc7e7ee844a1bc4dd89c332"} Sep 29 19:27:54 crc kubenswrapper[4741]: I0929 19:27:54.716064 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 19:27:55 crc kubenswrapper[4741]: I0929 19:27:55.242856 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 19:27:55 crc kubenswrapper[4741]: I0929 19:27:55.247014 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 19:27:55 crc kubenswrapper[4741]: I0929 19:27:55.578732 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:27:55 crc kubenswrapper[4741]: I0929 19:27:55.598258 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-677b8f6f88-c7kf7" podStartSLOduration=2.598238573 podStartE2EDuration="2.598238573s" podCreationTimestamp="2025-09-29 19:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:27:55.59309985 +0000 UTC m=+1117.240889192" watchObservedRunningTime="2025-09-29 19:27:55.598238573 +0000 UTC m=+1117.246027905" Sep 29 19:27:56 crc kubenswrapper[4741]: I0929 19:27:56.897563 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:27:56 crc kubenswrapper[4741]: I0929 19:27:56.950511 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-nt54p"] Sep 29 19:27:56 crc kubenswrapper[4741]: I0929 19:27:56.950907 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" containerName="dnsmasq-dns" containerID="cri-o://5b46b99038b674100da22bc5da4251f6cd7fb3da90cde9218f7543a7494d97b6" gracePeriod=10 Sep 29 19:27:57 crc kubenswrapper[4741]: I0929 19:27:57.599126 4741 generic.go:334] "Generic (PLEG): container finished" podID="00111435-439e-47b2-b8e8-08ee8fa40444" containerID="5b46b99038b674100da22bc5da4251f6cd7fb3da90cde9218f7543a7494d97b6" exitCode=0 Sep 29 19:27:57 crc kubenswrapper[4741]: I0929 19:27:57.599173 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" event={"ID":"00111435-439e-47b2-b8e8-08ee8fa40444","Type":"ContainerDied","Data":"5b46b99038b674100da22bc5da4251f6cd7fb3da90cde9218f7543a7494d97b6"} Sep 29 19:27:58 crc kubenswrapper[4741]: I0929 19:27:58.258286 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: connect: connection refused" Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.914252 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.929689 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-swift-storage-0\") pod \"00111435-439e-47b2-b8e8-08ee8fa40444\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.929862 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-svc\") pod \"00111435-439e-47b2-b8e8-08ee8fa40444\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.929937 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sglhw\" (UniqueName: \"kubernetes.io/projected/00111435-439e-47b2-b8e8-08ee8fa40444-kube-api-access-sglhw\") pod \"00111435-439e-47b2-b8e8-08ee8fa40444\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.929982 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-nb\") pod \"00111435-439e-47b2-b8e8-08ee8fa40444\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.930012 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-sb\") pod \"00111435-439e-47b2-b8e8-08ee8fa40444\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.930045 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-config\") pod \"00111435-439e-47b2-b8e8-08ee8fa40444\" (UID: \"00111435-439e-47b2-b8e8-08ee8fa40444\") " Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.971868 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00111435-439e-47b2-b8e8-08ee8fa40444-kube-api-access-sglhw" (OuterVolumeSpecName: "kube-api-access-sglhw") pod "00111435-439e-47b2-b8e8-08ee8fa40444" (UID: "00111435-439e-47b2-b8e8-08ee8fa40444"). InnerVolumeSpecName "kube-api-access-sglhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:00 crc kubenswrapper[4741]: I0929 19:28:00.992223 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00111435-439e-47b2-b8e8-08ee8fa40444" (UID: "00111435-439e-47b2-b8e8-08ee8fa40444"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.013798 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00111435-439e-47b2-b8e8-08ee8fa40444" (UID: "00111435-439e-47b2-b8e8-08ee8fa40444"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.031673 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sglhw\" (UniqueName: \"kubernetes.io/projected/00111435-439e-47b2-b8e8-08ee8fa40444-kube-api-access-sglhw\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.031697 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.031706 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.038564 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "00111435-439e-47b2-b8e8-08ee8fa40444" (UID: "00111435-439e-47b2-b8e8-08ee8fa40444"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.044833 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-config" (OuterVolumeSpecName: "config") pod "00111435-439e-47b2-b8e8-08ee8fa40444" (UID: "00111435-439e-47b2-b8e8-08ee8fa40444"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.049802 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00111435-439e-47b2-b8e8-08ee8fa40444" (UID: "00111435-439e-47b2-b8e8-08ee8fa40444"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.132796 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.132824 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.132836 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00111435-439e-47b2-b8e8-08ee8fa40444-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:01 crc kubenswrapper[4741]: E0929 19:28:01.289872 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.645337 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerStarted","Data":"3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9"} Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.645510 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="ceilometer-notification-agent" containerID="cri-o://5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d" gracePeriod=30 Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.645735 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.645830 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="proxy-httpd" containerID="cri-o://3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9" gracePeriod=30 Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.645923 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="sg-core" containerID="cri-o://7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437" gracePeriod=30 Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.664191 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" event={"ID":"00111435-439e-47b2-b8e8-08ee8fa40444","Type":"ContainerDied","Data":"2d8f8feda7fc0e483eca7c54a56ebdda056d4d71041f5a6648457e62b461c04d"} Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.664246 4741 scope.go:117] "RemoveContainer" containerID="5b46b99038b674100da22bc5da4251f6cd7fb3da90cde9218f7543a7494d97b6" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.664434 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-nt54p" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.688529 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-646p8" event={"ID":"2e91d309-4c23-4d57-ad3f-8eb7634b9b04","Type":"ContainerStarted","Data":"0e81cfe546352fd6f5ad379bfb7d977c1d18997906a11e352f7345acc20dfc48"} Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.711520 4741 scope.go:117] "RemoveContainer" containerID="8ccf185fed60f76c71590d2eb3b70147456c143c000e4ac7b9e5496f4ad963ab" Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.736954 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-nt54p"] Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.744610 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-nt54p"] Sep 29 19:28:01 crc kubenswrapper[4741]: I0929 19:28:01.750044 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-646p8" podStartSLOduration=2.7227906600000003 podStartE2EDuration="39.750033474s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="2025-09-29 19:27:23.935367379 +0000 UTC m=+1085.583156711" lastFinishedPulling="2025-09-29 19:28:00.962610193 +0000 UTC m=+1122.610399525" observedRunningTime="2025-09-29 19:28:01.715683878 +0000 UTC m=+1123.363473220" watchObservedRunningTime="2025-09-29 19:28:01.750033474 +0000 UTC m=+1123.397822806" Sep 29 19:28:02 crc kubenswrapper[4741]: I0929 19:28:02.700435 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gfpz8" event={"ID":"e172bee5-8543-4044-bf71-57e488d14fe4","Type":"ContainerStarted","Data":"bf70af070e1337a80f352e7891b4066ae9a350ffce11fe54650a0f5325a94c18"} Sep 29 19:28:02 crc kubenswrapper[4741]: I0929 19:28:02.703860 4741 generic.go:334] "Generic (PLEG): container finished" podID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerID="3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9" exitCode=0 Sep 29 19:28:02 crc kubenswrapper[4741]: I0929 19:28:02.704017 4741 generic.go:334] "Generic (PLEG): container finished" podID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerID="7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437" exitCode=2 Sep 29 19:28:02 crc kubenswrapper[4741]: I0929 19:28:02.703939 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerDied","Data":"3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9"} Sep 29 19:28:02 crc kubenswrapper[4741]: I0929 19:28:02.704285 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerDied","Data":"7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437"} Sep 29 19:28:02 crc kubenswrapper[4741]: I0929 19:28:02.728481 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-gfpz8" podStartSLOduration=3.331696216 podStartE2EDuration="40.728368973s" podCreationTimestamp="2025-09-29 19:27:22 +0000 UTC" firstStartedPulling="2025-09-29 19:27:23.561517236 +0000 UTC m=+1085.209306568" lastFinishedPulling="2025-09-29 19:28:00.958189993 +0000 UTC m=+1122.605979325" observedRunningTime="2025-09-29 19:28:02.718273894 +0000 UTC m=+1124.366063266" watchObservedRunningTime="2025-09-29 19:28:02.728368973 +0000 UTC m=+1124.376158345" Sep 29 19:28:03 crc kubenswrapper[4741]: I0929 19:28:03.100319 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" path="/var/lib/kubelet/pods/00111435-439e-47b2-b8e8-08ee8fa40444/volumes" Sep 29 19:28:03 crc kubenswrapper[4741]: I0929 19:28:03.716252 4741 generic.go:334] "Generic (PLEG): container finished" podID="2e91d309-4c23-4d57-ad3f-8eb7634b9b04" containerID="0e81cfe546352fd6f5ad379bfb7d977c1d18997906a11e352f7345acc20dfc48" exitCode=0 Sep 29 19:28:03 crc kubenswrapper[4741]: I0929 19:28:03.716328 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-646p8" event={"ID":"2e91d309-4c23-4d57-ad3f-8eb7634b9b04","Type":"ContainerDied","Data":"0e81cfe546352fd6f5ad379bfb7d977c1d18997906a11e352f7345acc20dfc48"} Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.118152 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-646p8" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.293954 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.302568 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-db-sync-config-data\") pod \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.302610 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-combined-ca-bundle\") pod \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.302707 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5fg7\" (UniqueName: \"kubernetes.io/projected/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-kube-api-access-q5fg7\") pod \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\" (UID: \"2e91d309-4c23-4d57-ad3f-8eb7634b9b04\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.308287 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2e91d309-4c23-4d57-ad3f-8eb7634b9b04" (UID: "2e91d309-4c23-4d57-ad3f-8eb7634b9b04"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.308803 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-kube-api-access-q5fg7" (OuterVolumeSpecName: "kube-api-access-q5fg7") pod "2e91d309-4c23-4d57-ad3f-8eb7634b9b04" (UID: "2e91d309-4c23-4d57-ad3f-8eb7634b9b04"). InnerVolumeSpecName "kube-api-access-q5fg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.346777 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e91d309-4c23-4d57-ad3f-8eb7634b9b04" (UID: "2e91d309-4c23-4d57-ad3f-8eb7634b9b04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.403738 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-config-data\") pod \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.403840 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-combined-ca-bundle\") pod \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.403910 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-scripts\") pod \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.403981 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxjtt\" (UniqueName: \"kubernetes.io/projected/34dc5bf3-42dd-47ec-ba42-7422df4dde51-kube-api-access-zxjtt\") pod \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.404013 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-run-httpd\") pod \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.404051 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-sg-core-conf-yaml\") pod \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.404080 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-log-httpd\") pod \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\" (UID: \"34dc5bf3-42dd-47ec-ba42-7422df4dde51\") " Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.404595 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5fg7\" (UniqueName: \"kubernetes.io/projected/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-kube-api-access-q5fg7\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.404619 4741 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.404630 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91d309-4c23-4d57-ad3f-8eb7634b9b04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.404663 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "34dc5bf3-42dd-47ec-ba42-7422df4dde51" (UID: "34dc5bf3-42dd-47ec-ba42-7422df4dde51"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.405234 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "34dc5bf3-42dd-47ec-ba42-7422df4dde51" (UID: "34dc5bf3-42dd-47ec-ba42-7422df4dde51"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.407375 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-scripts" (OuterVolumeSpecName: "scripts") pod "34dc5bf3-42dd-47ec-ba42-7422df4dde51" (UID: "34dc5bf3-42dd-47ec-ba42-7422df4dde51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.408457 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34dc5bf3-42dd-47ec-ba42-7422df4dde51-kube-api-access-zxjtt" (OuterVolumeSpecName: "kube-api-access-zxjtt") pod "34dc5bf3-42dd-47ec-ba42-7422df4dde51" (UID: "34dc5bf3-42dd-47ec-ba42-7422df4dde51"). InnerVolumeSpecName "kube-api-access-zxjtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.426057 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "34dc5bf3-42dd-47ec-ba42-7422df4dde51" (UID: "34dc5bf3-42dd-47ec-ba42-7422df4dde51"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.464003 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-config-data" (OuterVolumeSpecName: "config-data") pod "34dc5bf3-42dd-47ec-ba42-7422df4dde51" (UID: "34dc5bf3-42dd-47ec-ba42-7422df4dde51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.470731 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34dc5bf3-42dd-47ec-ba42-7422df4dde51" (UID: "34dc5bf3-42dd-47ec-ba42-7422df4dde51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.506423 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxjtt\" (UniqueName: \"kubernetes.io/projected/34dc5bf3-42dd-47ec-ba42-7422df4dde51-kube-api-access-zxjtt\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.506454 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.506465 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.506473 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34dc5bf3-42dd-47ec-ba42-7422df4dde51-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.506481 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.506489 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.506498 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34dc5bf3-42dd-47ec-ba42-7422df4dde51-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.738354 4741 generic.go:334] "Generic (PLEG): container finished" podID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerID="5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d" exitCode=0 Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.738469 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerDied","Data":"5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d"} Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.738525 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34dc5bf3-42dd-47ec-ba42-7422df4dde51","Type":"ContainerDied","Data":"bbe08b345bd3470436caafeeb50334a1f598e98ec9e674b5f2f5a87dbfd79645"} Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.738556 4741 scope.go:117] "RemoveContainer" containerID="3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.738956 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.741336 4741 generic.go:334] "Generic (PLEG): container finished" podID="e172bee5-8543-4044-bf71-57e488d14fe4" containerID="bf70af070e1337a80f352e7891b4066ae9a350ffce11fe54650a0f5325a94c18" exitCode=0 Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.741434 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gfpz8" event={"ID":"e172bee5-8543-4044-bf71-57e488d14fe4","Type":"ContainerDied","Data":"bf70af070e1337a80f352e7891b4066ae9a350ffce11fe54650a0f5325a94c18"} Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.748652 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-646p8" event={"ID":"2e91d309-4c23-4d57-ad3f-8eb7634b9b04","Type":"ContainerDied","Data":"2556098fc6e89af7c6c351ad10e0573f555efe8c133326fa21eefb11b9902401"} Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.748710 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2556098fc6e89af7c6c351ad10e0573f555efe8c133326fa21eefb11b9902401" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.748724 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-646p8" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.782904 4741 scope.go:117] "RemoveContainer" containerID="7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.832607 4741 scope.go:117] "RemoveContainer" containerID="5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.861588 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.871873 4741 scope.go:117] "RemoveContainer" containerID="3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.872342 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9\": container with ID starting with 3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9 not found: ID does not exist" containerID="3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.872382 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9"} err="failed to get container status \"3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9\": rpc error: code = NotFound desc = could not find container \"3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9\": container with ID starting with 3c39cfe43ee7bf897490fae091eba0654ab04ed836318e26b0c4f3c55701d3b9 not found: ID does not exist" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.872425 4741 scope.go:117] "RemoveContainer" containerID="7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.872761 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437\": container with ID starting with 7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437 not found: ID does not exist" containerID="7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.872811 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437"} err="failed to get container status \"7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437\": rpc error: code = NotFound desc = could not find container \"7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437\": container with ID starting with 7de1df112d573ed5054d27d6d3d10c0ad1555c006988d7d6dc93dd0cc5943437 not found: ID does not exist" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.872836 4741 scope.go:117] "RemoveContainer" containerID="5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.873078 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d\": container with ID starting with 5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d not found: ID does not exist" containerID="5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.873100 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d"} err="failed to get container status \"5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d\": rpc error: code = NotFound desc = could not find container \"5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d\": container with ID starting with 5aafad69f44f92eb34d365e00028fe94c1d05cf2d4440e4a96f5df378605583d not found: ID does not exist" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.873925 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881142 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.881559 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" containerName="dnsmasq-dns" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881582 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" containerName="dnsmasq-dns" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.881604 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="proxy-httpd" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881613 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="proxy-httpd" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.881626 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="ceilometer-notification-agent" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881638 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="ceilometer-notification-agent" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.881659 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" containerName="init" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881669 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" containerName="init" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.881688 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e91d309-4c23-4d57-ad3f-8eb7634b9b04" containerName="barbican-db-sync" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881696 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e91d309-4c23-4d57-ad3f-8eb7634b9b04" containerName="barbican-db-sync" Sep 29 19:28:05 crc kubenswrapper[4741]: E0929 19:28:05.881712 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="sg-core" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881721 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="sg-core" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881942 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="00111435-439e-47b2-b8e8-08ee8fa40444" containerName="dnsmasq-dns" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881961 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="sg-core" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.881984 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e91d309-4c23-4d57-ad3f-8eb7634b9b04" containerName="barbican-db-sync" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.882010 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="proxy-httpd" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.882024 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" containerName="ceilometer-notification-agent" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.883976 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.888223 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.888314 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 19:28:05 crc kubenswrapper[4741]: I0929 19:28:05.890044 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.003603 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-8567fd5894-dwdtp"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.005860 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.012107 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x729m" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.012459 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.012620 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.024148 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-log-httpd\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.024222 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.024262 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-config-data\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.024295 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-run-httpd\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.024312 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-scripts\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.024425 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.024456 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2wpx\" (UniqueName: \"kubernetes.io/projected/3390cf5f-a926-4ea3-9fae-d560cd99789a-kube-api-access-d2wpx\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.028075 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5c96956bc-7qhgv"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.029615 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.033725 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.039651 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8567fd5894-dwdtp"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.053533 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c96956bc-7qhgv"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.103135 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sxmfb"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.105213 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.116850 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sxmfb"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125416 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125487 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125530 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2wpx\" (UniqueName: \"kubernetes.io/projected/3390cf5f-a926-4ea3-9fae-d560cd99789a-kube-api-access-d2wpx\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125559 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-logs\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125581 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-log-httpd\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125605 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125630 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-config-data\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125657 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-run-httpd\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125674 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-scripts\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125690 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data-custom\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125736 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9txfg\" (UniqueName: \"kubernetes.io/projected/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-kube-api-access-9txfg\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.125757 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-combined-ca-bundle\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.127080 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-run-httpd\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.127333 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-log-httpd\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.131478 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.134822 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-config-data\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.138579 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-scripts\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.155159 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.158026 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2wpx\" (UniqueName: \"kubernetes.io/projected/3390cf5f-a926-4ea3-9fae-d560cd99789a-kube-api-access-d2wpx\") pod \"ceilometer-0\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.205498 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227568 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-combined-ca-bundle\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227665 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-svc\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227723 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227743 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227781 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227800 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-logs\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227815 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-config\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227843 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227857 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-logs\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.227912 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data-custom\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.228012 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data-custom\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.228031 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-925md\" (UniqueName: \"kubernetes.io/projected/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-kube-api-access-925md\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.228111 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srqc2\" (UniqueName: \"kubernetes.io/projected/d9284e2e-01fa-4a1a-9f31-bc47191028df-kube-api-access-srqc2\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.228144 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9txfg\" (UniqueName: \"kubernetes.io/projected/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-kube-api-access-9txfg\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.228170 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-combined-ca-bundle\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.229704 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-logs\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.234448 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data-custom\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.234983 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-combined-ca-bundle\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.240573 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.260000 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9txfg\" (UniqueName: \"kubernetes.io/projected/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-kube-api-access-9txfg\") pod \"barbican-keystone-listener-8567fd5894-dwdtp\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.269485 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5c8494f4cd-9bsk9"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.271494 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.279934 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.300019 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c8494f4cd-9bsk9"] Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.329659 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.330680 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-svc\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.330613 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.330765 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.331341 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-svc\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.331475 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.331506 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-logs\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.331523 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-config\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.331558 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.331624 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.331836 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-logs\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.332468 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.332547 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.333237 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data-custom\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.333301 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-925md\" (UniqueName: \"kubernetes.io/projected/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-kube-api-access-925md\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.333363 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srqc2\" (UniqueName: \"kubernetes.io/projected/d9284e2e-01fa-4a1a-9f31-bc47191028df-kube-api-access-srqc2\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.333472 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-config\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.333476 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-combined-ca-bundle\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.341206 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data-custom\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.341795 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-combined-ca-bundle\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.343512 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.353372 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-925md\" (UniqueName: \"kubernetes.io/projected/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-kube-api-access-925md\") pod \"barbican-worker-5c96956bc-7qhgv\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.354266 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srqc2\" (UniqueName: \"kubernetes.io/projected/d9284e2e-01fa-4a1a-9f31-bc47191028df-kube-api-access-srqc2\") pod \"dnsmasq-dns-85ff748b95-sxmfb\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.362990 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.427118 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.434865 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-combined-ca-bundle\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.434965 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.435022 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e918ae11-61d6-4766-b731-dc3b9f9c6abc-logs\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.435045 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pvwg\" (UniqueName: \"kubernetes.io/projected/e918ae11-61d6-4766-b731-dc3b9f9c6abc-kube-api-access-7pvwg\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.435067 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data-custom\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.537302 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.537411 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e918ae11-61d6-4766-b731-dc3b9f9c6abc-logs\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.537463 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pvwg\" (UniqueName: \"kubernetes.io/projected/e918ae11-61d6-4766-b731-dc3b9f9c6abc-kube-api-access-7pvwg\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.537585 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data-custom\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.537640 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-combined-ca-bundle\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.541093 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e918ae11-61d6-4766-b731-dc3b9f9c6abc-logs\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.562104 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-combined-ca-bundle\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.562666 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data-custom\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.563749 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.578528 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pvwg\" (UniqueName: \"kubernetes.io/projected/e918ae11-61d6-4766-b731-dc3b9f9c6abc-kube-api-access-7pvwg\") pod \"barbican-api-5c8494f4cd-9bsk9\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.647622 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:06 crc kubenswrapper[4741]: I0929 19:28:06.810731 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:06 crc kubenswrapper[4741]: W0929 19:28:06.821022 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3390cf5f_a926_4ea3_9fae_d560cd99789a.slice/crio-66f4828960adaec5a0e0c33c00725b6f6c639909f811bd688f25a56104d89fce WatchSource:0}: Error finding container 66f4828960adaec5a0e0c33c00725b6f6c639909f811bd688f25a56104d89fce: Status 404 returned error can't find the container with id 66f4828960adaec5a0e0c33c00725b6f6c639909f811bd688f25a56104d89fce Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.006274 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c96956bc-7qhgv"] Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.035625 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8567fd5894-dwdtp"] Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.097179 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34dc5bf3-42dd-47ec-ba42-7422df4dde51" path="/var/lib/kubelet/pods/34dc5bf3-42dd-47ec-ba42-7422df4dde51/volumes" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.153935 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sxmfb"] Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.242168 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c8494f4cd-9bsk9"] Sep 29 19:28:07 crc kubenswrapper[4741]: W0929 19:28:07.262328 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode918ae11_61d6_4766_b731_dc3b9f9c6abc.slice/crio-20670d470185d459ac22f7e767252d4aaa388b331ed1697f0c54ddd48cffae85 WatchSource:0}: Error finding container 20670d470185d459ac22f7e767252d4aaa388b331ed1697f0c54ddd48cffae85: Status 404 returned error can't find the container with id 20670d470185d459ac22f7e767252d4aaa388b331ed1697f0c54ddd48cffae85 Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.350488 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.455534 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-config-data\") pod \"e172bee5-8543-4044-bf71-57e488d14fe4\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.455596 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-db-sync-config-data\") pod \"e172bee5-8543-4044-bf71-57e488d14fe4\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.455637 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-combined-ca-bundle\") pod \"e172bee5-8543-4044-bf71-57e488d14fe4\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.455698 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-scripts\") pod \"e172bee5-8543-4044-bf71-57e488d14fe4\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.455728 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e172bee5-8543-4044-bf71-57e488d14fe4-etc-machine-id\") pod \"e172bee5-8543-4044-bf71-57e488d14fe4\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.455825 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67s8w\" (UniqueName: \"kubernetes.io/projected/e172bee5-8543-4044-bf71-57e488d14fe4-kube-api-access-67s8w\") pod \"e172bee5-8543-4044-bf71-57e488d14fe4\" (UID: \"e172bee5-8543-4044-bf71-57e488d14fe4\") " Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.456161 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e172bee5-8543-4044-bf71-57e488d14fe4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e172bee5-8543-4044-bf71-57e488d14fe4" (UID: "e172bee5-8543-4044-bf71-57e488d14fe4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.456441 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e172bee5-8543-4044-bf71-57e488d14fe4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.462355 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e172bee5-8543-4044-bf71-57e488d14fe4" (UID: "e172bee5-8543-4044-bf71-57e488d14fe4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.462578 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e172bee5-8543-4044-bf71-57e488d14fe4-kube-api-access-67s8w" (OuterVolumeSpecName: "kube-api-access-67s8w") pod "e172bee5-8543-4044-bf71-57e488d14fe4" (UID: "e172bee5-8543-4044-bf71-57e488d14fe4"). InnerVolumeSpecName "kube-api-access-67s8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.463525 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-scripts" (OuterVolumeSpecName: "scripts") pod "e172bee5-8543-4044-bf71-57e488d14fe4" (UID: "e172bee5-8543-4044-bf71-57e488d14fe4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.493539 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e172bee5-8543-4044-bf71-57e488d14fe4" (UID: "e172bee5-8543-4044-bf71-57e488d14fe4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.510379 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-config-data" (OuterVolumeSpecName: "config-data") pod "e172bee5-8543-4044-bf71-57e488d14fe4" (UID: "e172bee5-8543-4044-bf71-57e488d14fe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.557806 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.557845 4741 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.557857 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.557867 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e172bee5-8543-4044-bf71-57e488d14fe4-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.557877 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67s8w\" (UniqueName: \"kubernetes.io/projected/e172bee5-8543-4044-bf71-57e488d14fe4-kube-api-access-67s8w\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.775128 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c8494f4cd-9bsk9" event={"ID":"e918ae11-61d6-4766-b731-dc3b9f9c6abc","Type":"ContainerStarted","Data":"75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.775217 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c8494f4cd-9bsk9" event={"ID":"e918ae11-61d6-4766-b731-dc3b9f9c6abc","Type":"ContainerStarted","Data":"7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.775233 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c8494f4cd-9bsk9" event={"ID":"e918ae11-61d6-4766-b731-dc3b9f9c6abc","Type":"ContainerStarted","Data":"20670d470185d459ac22f7e767252d4aaa388b331ed1697f0c54ddd48cffae85"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.775663 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.775723 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.789718 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerStarted","Data":"d04ab1347974022043dd25c8315eebd22bbe9958a62469dfe66ccd082f29ace1"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.789800 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerStarted","Data":"66f4828960adaec5a0e0c33c00725b6f6c639909f811bd688f25a56104d89fce"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.792661 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c96956bc-7qhgv" event={"ID":"d3e3a43d-5ff8-491d-916c-3c0b9e31f223","Type":"ContainerStarted","Data":"f524324ad5b249dd4c548da49178acfd62c8196cee5665981e9ad768742d315b"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.802935 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gfpz8" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.804005 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gfpz8" event={"ID":"e172bee5-8543-4044-bf71-57e488d14fe4","Type":"ContainerDied","Data":"4ed66f64565d1f1c11d4dc81367805d6d8aa89947eef157cd3134336fcb5aea7"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.804091 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ed66f64565d1f1c11d4dc81367805d6d8aa89947eef157cd3134336fcb5aea7" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.809971 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5c8494f4cd-9bsk9" podStartSLOduration=1.809951911 podStartE2EDuration="1.809951911s" podCreationTimestamp="2025-09-29 19:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:07.795347609 +0000 UTC m=+1129.443136941" watchObservedRunningTime="2025-09-29 19:28:07.809951911 +0000 UTC m=+1129.457741243" Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.814397 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" event={"ID":"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d","Type":"ContainerStarted","Data":"b42500dea64ea4de76dc6f41aa6240da449020927e24ac47aca39f94f9c04e59"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.817084 4741 generic.go:334] "Generic (PLEG): container finished" podID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerID="27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc" exitCode=0 Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.817133 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" event={"ID":"d9284e2e-01fa-4a1a-9f31-bc47191028df","Type":"ContainerDied","Data":"27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc"} Sep 29 19:28:07 crc kubenswrapper[4741]: I0929 19:28:07.817177 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" event={"ID":"d9284e2e-01fa-4a1a-9f31-bc47191028df","Type":"ContainerStarted","Data":"01fe8e1419b80f5cb9cfd2528d948c128abc569e539deb5c859ae51dcb9b6818"} Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.002199 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:08 crc kubenswrapper[4741]: E0929 19:28:08.002857 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e172bee5-8543-4044-bf71-57e488d14fe4" containerName="cinder-db-sync" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.002870 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e172bee5-8543-4044-bf71-57e488d14fe4" containerName="cinder-db-sync" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.003061 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e172bee5-8543-4044-bf71-57e488d14fe4" containerName="cinder-db-sync" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.007079 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.010449 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.028297 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.028766 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.028913 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-vb8vq" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.028997 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.122373 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sxmfb"] Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.133557 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-m7rm4"] Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.135736 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.159400 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-m7rm4"] Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.175099 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b970fb5-e6f0-44e0-9dd1-6391661533ed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.175168 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.175190 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qhz7\" (UniqueName: \"kubernetes.io/projected/8b970fb5-e6f0-44e0-9dd1-6391661533ed-kube-api-access-2qhz7\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.175274 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.175288 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-scripts\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.175317 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.194687 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.201379 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.204578 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.211454 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.276590 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.276907 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277065 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b970fb5-e6f0-44e0-9dd1-6391661533ed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277173 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277299 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277373 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b970fb5-e6f0-44e0-9dd1-6391661533ed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277535 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qhz7\" (UniqueName: \"kubernetes.io/projected/8b970fb5-e6f0-44e0-9dd1-6391661533ed-kube-api-access-2qhz7\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277672 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277815 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-config\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.277933 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxnj2\" (UniqueName: \"kubernetes.io/projected/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-kube-api-access-jxnj2\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.278050 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.278153 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-scripts\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.278258 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.281744 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.282148 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.294410 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-scripts\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.295493 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.298526 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qhz7\" (UniqueName: \"kubernetes.io/projected/8b970fb5-e6f0-44e0-9dd1-6391661533ed-kube-api-access-2qhz7\") pod \"cinder-scheduler-0\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.345281 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.379813 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.379881 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.379909 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48hqj\" (UniqueName: \"kubernetes.io/projected/1516035e-4dd8-48ab-945a-09f7bee4bf22-kube-api-access-48hqj\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.379929 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data-custom\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.379955 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1516035e-4dd8-48ab-945a-09f7bee4bf22-logs\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.379995 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-config\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380028 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxnj2\" (UniqueName: \"kubernetes.io/projected/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-kube-api-access-jxnj2\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380057 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-scripts\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380083 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380119 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380146 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1516035e-4dd8-48ab-945a-09f7bee4bf22-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380175 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380195 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.380903 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.381070 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.381533 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-config\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.382625 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.383068 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.397122 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxnj2\" (UniqueName: \"kubernetes.io/projected/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-kube-api-access-jxnj2\") pod \"dnsmasq-dns-5c9776ccc5-m7rm4\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.462994 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482109 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1516035e-4dd8-48ab-945a-09f7bee4bf22-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482168 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482212 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1516035e-4dd8-48ab-945a-09f7bee4bf22-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482216 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482305 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48hqj\" (UniqueName: \"kubernetes.io/projected/1516035e-4dd8-48ab-945a-09f7bee4bf22-kube-api-access-48hqj\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482324 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data-custom\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482349 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1516035e-4dd8-48ab-945a-09f7bee4bf22-logs\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482450 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-scripts\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.482768 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1516035e-4dd8-48ab-945a-09f7bee4bf22-logs\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.486478 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.487738 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-scripts\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.498536 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data-custom\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.499658 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.501757 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48hqj\" (UniqueName: \"kubernetes.io/projected/1516035e-4dd8-48ab-945a-09f7bee4bf22-kube-api-access-48hqj\") pod \"cinder-api-0\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " pod="openstack/cinder-api-0" Sep 29 19:28:08 crc kubenswrapper[4741]: I0929 19:28:08.526547 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.531958 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.553865 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-m7rm4"] Sep 29 19:28:09 crc kubenswrapper[4741]: W0929 19:28:09.561869 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d404bf3_e8d2_4010_8c59_bc6a6e0c1fb0.slice/crio-3256c8ad423d65ea473b4f3da8441b6bf99d3ef8bb5522ebcd79ca61b0561003 WatchSource:0}: Error finding container 3256c8ad423d65ea473b4f3da8441b6bf99d3ef8bb5522ebcd79ca61b0561003: Status 404 returned error can't find the container with id 3256c8ad423d65ea473b4f3da8441b6bf99d3ef8bb5522ebcd79ca61b0561003 Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.673462 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:09 crc kubenswrapper[4741]: W0929 19:28:09.679262 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b970fb5_e6f0_44e0_9dd1_6391661533ed.slice/crio-32e0b9115f10f276b97f9ba483d6e0b1f197463f2f2757a4caad06a4e9883eff WatchSource:0}: Error finding container 32e0b9115f10f276b97f9ba483d6e0b1f197463f2f2757a4caad06a4e9883eff: Status 404 returned error can't find the container with id 32e0b9115f10f276b97f9ba483d6e0b1f197463f2f2757a4caad06a4e9883eff Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.841149 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8b970fb5-e6f0-44e0-9dd1-6391661533ed","Type":"ContainerStarted","Data":"32e0b9115f10f276b97f9ba483d6e0b1f197463f2f2757a4caad06a4e9883eff"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.844587 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" event={"ID":"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d","Type":"ContainerStarted","Data":"4b30dc6f2664d0596e57ddb9855aca996f10f7caab7b0df97ee564ddcc5bc232"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.844615 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" event={"ID":"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d","Type":"ContainerStarted","Data":"fc163c8b456c4eae3276be79b996f9a990070e39bb81a710df286eea59af6f36"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.851176 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" podUID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerName="dnsmasq-dns" containerID="cri-o://43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7" gracePeriod=10 Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.851166 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" event={"ID":"d9284e2e-01fa-4a1a-9f31-bc47191028df","Type":"ContainerStarted","Data":"43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.851288 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.854140 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerStarted","Data":"698002fd90572118b08152d0ec223b9498eef909ee7b1674d236a8b8b8fc3f79"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.856168 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1516035e-4dd8-48ab-945a-09f7bee4bf22","Type":"ContainerStarted","Data":"6ca33e3a8b281d6e9780b0ee013dd9c8a83be986a2e3835d6ab9a4ddecfbd0cc"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.862515 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c96956bc-7qhgv" event={"ID":"d3e3a43d-5ff8-491d-916c-3c0b9e31f223","Type":"ContainerStarted","Data":"5d7c9dcdb6e6631c524f68a99886213580f717eb067525088f01ff7e8451b3e4"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.862580 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c96956bc-7qhgv" event={"ID":"d3e3a43d-5ff8-491d-916c-3c0b9e31f223","Type":"ContainerStarted","Data":"f4540013c871a62255548905e7c7a5f8f143d8e4617fe18f7c6f7bdbd529b257"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.867804 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" event={"ID":"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0","Type":"ContainerStarted","Data":"8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.867874 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" event={"ID":"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0","Type":"ContainerStarted","Data":"3256c8ad423d65ea473b4f3da8441b6bf99d3ef8bb5522ebcd79ca61b0561003"} Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.894974 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" podStartSLOduration=2.8709587 podStartE2EDuration="4.894957206s" podCreationTimestamp="2025-09-29 19:28:05 +0000 UTC" firstStartedPulling="2025-09-29 19:28:07.003990434 +0000 UTC m=+1128.651779766" lastFinishedPulling="2025-09-29 19:28:09.02798894 +0000 UTC m=+1130.675778272" observedRunningTime="2025-09-29 19:28:09.861295711 +0000 UTC m=+1131.509085043" watchObservedRunningTime="2025-09-29 19:28:09.894957206 +0000 UTC m=+1131.542746538" Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.900323 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" podStartSLOduration=3.900273314 podStartE2EDuration="3.900273314s" podCreationTimestamp="2025-09-29 19:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:09.881121458 +0000 UTC m=+1131.528910790" watchObservedRunningTime="2025-09-29 19:28:09.900273314 +0000 UTC m=+1131.548062646" Sep 29 19:28:09 crc kubenswrapper[4741]: I0929 19:28:09.931383 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5c96956bc-7qhgv" podStartSLOduration=2.960611145 podStartE2EDuration="4.931358837s" podCreationTimestamp="2025-09-29 19:28:05 +0000 UTC" firstStartedPulling="2025-09-29 19:28:07.004024375 +0000 UTC m=+1128.651813707" lastFinishedPulling="2025-09-29 19:28:08.974772067 +0000 UTC m=+1130.622561399" observedRunningTime="2025-09-29 19:28:09.916639071 +0000 UTC m=+1131.564428403" watchObservedRunningTime="2025-09-29 19:28:09.931358837 +0000 UTC m=+1131.579148169" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.642331 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.652603 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-swift-storage-0\") pod \"d9284e2e-01fa-4a1a-9f31-bc47191028df\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.652671 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-sb\") pod \"d9284e2e-01fa-4a1a-9f31-bc47191028df\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.652703 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-nb\") pod \"d9284e2e-01fa-4a1a-9f31-bc47191028df\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.652749 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srqc2\" (UniqueName: \"kubernetes.io/projected/d9284e2e-01fa-4a1a-9f31-bc47191028df-kube-api-access-srqc2\") pod \"d9284e2e-01fa-4a1a-9f31-bc47191028df\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.653686 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-config\") pod \"d9284e2e-01fa-4a1a-9f31-bc47191028df\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.653721 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-svc\") pod \"d9284e2e-01fa-4a1a-9f31-bc47191028df\" (UID: \"d9284e2e-01fa-4a1a-9f31-bc47191028df\") " Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.662233 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9284e2e-01fa-4a1a-9f31-bc47191028df-kube-api-access-srqc2" (OuterVolumeSpecName: "kube-api-access-srqc2") pod "d9284e2e-01fa-4a1a-9f31-bc47191028df" (UID: "d9284e2e-01fa-4a1a-9f31-bc47191028df"). InnerVolumeSpecName "kube-api-access-srqc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.710727 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d9284e2e-01fa-4a1a-9f31-bc47191028df" (UID: "d9284e2e-01fa-4a1a-9f31-bc47191028df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.726697 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d9284e2e-01fa-4a1a-9f31-bc47191028df" (UID: "d9284e2e-01fa-4a1a-9f31-bc47191028df"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.726870 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d9284e2e-01fa-4a1a-9f31-bc47191028df" (UID: "d9284e2e-01fa-4a1a-9f31-bc47191028df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.749538 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d9284e2e-01fa-4a1a-9f31-bc47191028df" (UID: "d9284e2e-01fa-4a1a-9f31-bc47191028df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.755698 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.755729 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.755742 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srqc2\" (UniqueName: \"kubernetes.io/projected/d9284e2e-01fa-4a1a-9f31-bc47191028df-kube-api-access-srqc2\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.755755 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.755764 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.774878 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-config" (OuterVolumeSpecName: "config") pod "d9284e2e-01fa-4a1a-9f31-bc47191028df" (UID: "d9284e2e-01fa-4a1a-9f31-bc47191028df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.858375 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9284e2e-01fa-4a1a-9f31-bc47191028df-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.882327 4741 generic.go:334] "Generic (PLEG): container finished" podID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerID="43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7" exitCode=0 Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.882414 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" event={"ID":"d9284e2e-01fa-4a1a-9f31-bc47191028df","Type":"ContainerDied","Data":"43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7"} Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.882447 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" event={"ID":"d9284e2e-01fa-4a1a-9f31-bc47191028df","Type":"ContainerDied","Data":"01fe8e1419b80f5cb9cfd2528d948c128abc569e539deb5c859ae51dcb9b6818"} Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.882467 4741 scope.go:117] "RemoveContainer" containerID="43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.882605 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sxmfb" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.893695 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerStarted","Data":"1c0079aa744f1248402f2e7a77e71d07d8e514fd609bf0af6d642f6a7997bb69"} Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.898867 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1516035e-4dd8-48ab-945a-09f7bee4bf22","Type":"ContainerStarted","Data":"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7"} Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.902035 4741 generic.go:334] "Generic (PLEG): container finished" podID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerID="8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb" exitCode=0 Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.902296 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" event={"ID":"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0","Type":"ContainerDied","Data":"8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb"} Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.902347 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" event={"ID":"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0","Type":"ContainerStarted","Data":"6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde"} Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.903059 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.917588 4741 scope.go:117] "RemoveContainer" containerID="27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.917711 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sxmfb"] Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.924504 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sxmfb"] Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.932454 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" podStartSLOduration=2.932438674 podStartE2EDuration="2.932438674s" podCreationTimestamp="2025-09-29 19:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:10.930067189 +0000 UTC m=+1132.577856541" watchObservedRunningTime="2025-09-29 19:28:10.932438674 +0000 UTC m=+1132.580228006" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.948894 4741 scope.go:117] "RemoveContainer" containerID="43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7" Sep 29 19:28:10 crc kubenswrapper[4741]: E0929 19:28:10.949534 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7\": container with ID starting with 43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7 not found: ID does not exist" containerID="43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.949568 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7"} err="failed to get container status \"43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7\": rpc error: code = NotFound desc = could not find container \"43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7\": container with ID starting with 43a1a7ef0feb7afecb2f22229f53784bced74ff5bdf9f2d6dc765d9df26bc1c7 not found: ID does not exist" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.949593 4741 scope.go:117] "RemoveContainer" containerID="27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc" Sep 29 19:28:10 crc kubenswrapper[4741]: E0929 19:28:10.949875 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc\": container with ID starting with 27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc not found: ID does not exist" containerID="27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc" Sep 29 19:28:10 crc kubenswrapper[4741]: I0929 19:28:10.949913 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc"} err="failed to get container status \"27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc\": rpc error: code = NotFound desc = could not find container \"27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc\": container with ID starting with 27c747a802485e7c4a3102c7d6d0a956e9a07078954355bfb827929956ce38bc not found: ID does not exist" Sep 29 19:28:11 crc kubenswrapper[4741]: I0929 19:28:11.127735 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9284e2e-01fa-4a1a-9f31-bc47191028df" path="/var/lib/kubelet/pods/d9284e2e-01fa-4a1a-9f31-bc47191028df/volumes" Sep 29 19:28:11 crc kubenswrapper[4741]: I0929 19:28:11.925264 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8b970fb5-e6f0-44e0-9dd1-6391661533ed","Type":"ContainerStarted","Data":"fc92aecb2dfb9850b26acbeeeaeffacdb85da5b76c5b3b7b8969e785ccce435e"} Sep 29 19:28:11 crc kubenswrapper[4741]: I0929 19:28:11.932949 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1516035e-4dd8-48ab-945a-09f7bee4bf22","Type":"ContainerStarted","Data":"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695"} Sep 29 19:28:11 crc kubenswrapper[4741]: I0929 19:28:11.962409 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.962362405 podStartE2EDuration="3.962362405s" podCreationTimestamp="2025-09-29 19:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:11.955636752 +0000 UTC m=+1133.603426124" watchObservedRunningTime="2025-09-29 19:28:11.962362405 +0000 UTC m=+1133.610151757" Sep 29 19:28:12 crc kubenswrapper[4741]: I0929 19:28:12.531074 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:12 crc kubenswrapper[4741]: I0929 19:28:12.945547 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerStarted","Data":"03c697990d9bf6337f6a018465b8ba174c224f446de542677af9f64c31f0c004"} Sep 29 19:28:12 crc kubenswrapper[4741]: I0929 19:28:12.945697 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 19:28:12 crc kubenswrapper[4741]: I0929 19:28:12.947597 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8b970fb5-e6f0-44e0-9dd1-6391661533ed","Type":"ContainerStarted","Data":"e668bdb14dae689ea32169dbee26af2d538bda4ff099cd9c2757e3a67cdabfb8"} Sep 29 19:28:12 crc kubenswrapper[4741]: I0929 19:28:12.947776 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.005432 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-659f5d9f44-j8zgh"] Sep 29 19:28:13 crc kubenswrapper[4741]: E0929 19:28:13.005764 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerName="dnsmasq-dns" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.005779 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerName="dnsmasq-dns" Sep 29 19:28:13 crc kubenswrapper[4741]: E0929 19:28:13.005814 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerName="init" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.005820 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerName="init" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.005986 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9284e2e-01fa-4a1a-9f31-bc47191028df" containerName="dnsmasq-dns" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.006867 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: W0929 19:28:13.009659 4741 reflector.go:561] object-"openstack"/"cert-barbican-internal-svc": failed to list *v1.Secret: secrets "cert-barbican-internal-svc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Sep 29 19:28:13 crc kubenswrapper[4741]: E0929 19:28:13.009700 4741 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-barbican-internal-svc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-barbican-internal-svc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 29 19:28:13 crc kubenswrapper[4741]: W0929 19:28:13.009739 4741 reflector.go:561] object-"openstack"/"cert-barbican-public-svc": failed to list *v1.Secret: secrets "cert-barbican-public-svc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Sep 29 19:28:13 crc kubenswrapper[4741]: E0929 19:28:13.009749 4741 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-barbican-public-svc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-barbican-public-svc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.038726 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-659f5d9f44-j8zgh"] Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.055020 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.763134511 podStartE2EDuration="8.055004868s" podCreationTimestamp="2025-09-29 19:28:05 +0000 UTC" firstStartedPulling="2025-09-29 19:28:06.825063125 +0000 UTC m=+1128.472852457" lastFinishedPulling="2025-09-29 19:28:12.116933482 +0000 UTC m=+1133.764722814" observedRunningTime="2025-09-29 19:28:13.034347205 +0000 UTC m=+1134.682136537" watchObservedRunningTime="2025-09-29 19:28:13.055004868 +0000 UTC m=+1134.702794200" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.059850 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.106759842 podStartE2EDuration="6.059835921s" podCreationTimestamp="2025-09-29 19:28:07 +0000 UTC" firstStartedPulling="2025-09-29 19:28:09.680637909 +0000 UTC m=+1131.328427241" lastFinishedPulling="2025-09-29 19:28:10.633713988 +0000 UTC m=+1132.281503320" observedRunningTime="2025-09-29 19:28:13.052153937 +0000 UTC m=+1134.699943269" watchObservedRunningTime="2025-09-29 19:28:13.059835921 +0000 UTC m=+1134.707625253" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.094673 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.095715 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data-custom\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.096294 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-logs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.096318 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldntx\" (UniqueName: \"kubernetes.io/projected/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-kube-api-access-ldntx\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.096550 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-combined-ca-bundle\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.098291 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-internal-tls-certs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.098522 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-public-tls-certs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.200800 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-logs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.200844 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldntx\" (UniqueName: \"kubernetes.io/projected/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-kube-api-access-ldntx\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.200880 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-combined-ca-bundle\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.200922 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-internal-tls-certs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.200976 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-public-tls-certs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.201016 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.201054 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data-custom\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.201239 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-logs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.210220 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-combined-ca-bundle\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.220075 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data-custom\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.228075 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldntx\" (UniqueName: \"kubernetes.io/projected/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-kube-api-access-ldntx\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.233400 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.345620 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.961316 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api-log" containerID="cri-o://9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7" gracePeriod=30 Sep 29 19:28:13 crc kubenswrapper[4741]: I0929 19:28:13.961383 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api" containerID="cri-o://707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695" gracePeriod=30 Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.002953 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.016920 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-public-tls-certs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.184980 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.196167 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-internal-tls-certs\") pod \"barbican-api-659f5d9f44-j8zgh\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.220900 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.650951 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.778856 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-659f5d9f44-j8zgh"] Sep 29 19:28:14 crc kubenswrapper[4741]: W0929 19:28:14.782892 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58cfa8a3_5067_40ee_8e23_803a52d6b1b1.slice/crio-dab1bbdc6afa761fe9a849e8ae5ff2eb85c04e27d9ed56440a407f6e2e4c7bf5 WatchSource:0}: Error finding container dab1bbdc6afa761fe9a849e8ae5ff2eb85c04e27d9ed56440a407f6e2e4c7bf5: Status 404 returned error can't find the container with id dab1bbdc6afa761fe9a849e8ae5ff2eb85c04e27d9ed56440a407f6e2e4c7bf5 Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.831112 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-scripts\") pod \"1516035e-4dd8-48ab-945a-09f7bee4bf22\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.831167 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1516035e-4dd8-48ab-945a-09f7bee4bf22-logs\") pod \"1516035e-4dd8-48ab-945a-09f7bee4bf22\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.831326 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-combined-ca-bundle\") pod \"1516035e-4dd8-48ab-945a-09f7bee4bf22\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.831369 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48hqj\" (UniqueName: \"kubernetes.io/projected/1516035e-4dd8-48ab-945a-09f7bee4bf22-kube-api-access-48hqj\") pod \"1516035e-4dd8-48ab-945a-09f7bee4bf22\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.831426 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1516035e-4dd8-48ab-945a-09f7bee4bf22-etc-machine-id\") pod \"1516035e-4dd8-48ab-945a-09f7bee4bf22\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.831444 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data-custom\") pod \"1516035e-4dd8-48ab-945a-09f7bee4bf22\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.831483 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data\") pod \"1516035e-4dd8-48ab-945a-09f7bee4bf22\" (UID: \"1516035e-4dd8-48ab-945a-09f7bee4bf22\") " Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.833117 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1516035e-4dd8-48ab-945a-09f7bee4bf22-logs" (OuterVolumeSpecName: "logs") pod "1516035e-4dd8-48ab-945a-09f7bee4bf22" (UID: "1516035e-4dd8-48ab-945a-09f7bee4bf22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.833963 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1516035e-4dd8-48ab-945a-09f7bee4bf22-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1516035e-4dd8-48ab-945a-09f7bee4bf22" (UID: "1516035e-4dd8-48ab-945a-09f7bee4bf22"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.837799 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-scripts" (OuterVolumeSpecName: "scripts") pod "1516035e-4dd8-48ab-945a-09f7bee4bf22" (UID: "1516035e-4dd8-48ab-945a-09f7bee4bf22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.840484 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1516035e-4dd8-48ab-945a-09f7bee4bf22-kube-api-access-48hqj" (OuterVolumeSpecName: "kube-api-access-48hqj") pod "1516035e-4dd8-48ab-945a-09f7bee4bf22" (UID: "1516035e-4dd8-48ab-945a-09f7bee4bf22"). InnerVolumeSpecName "kube-api-access-48hqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.840587 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1516035e-4dd8-48ab-945a-09f7bee4bf22" (UID: "1516035e-4dd8-48ab-945a-09f7bee4bf22"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.878958 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1516035e-4dd8-48ab-945a-09f7bee4bf22" (UID: "1516035e-4dd8-48ab-945a-09f7bee4bf22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.930062 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data" (OuterVolumeSpecName: "config-data") pod "1516035e-4dd8-48ab-945a-09f7bee4bf22" (UID: "1516035e-4dd8-48ab-945a-09f7bee4bf22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.933761 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.933805 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1516035e-4dd8-48ab-945a-09f7bee4bf22-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.933827 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.933846 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48hqj\" (UniqueName: \"kubernetes.io/projected/1516035e-4dd8-48ab-945a-09f7bee4bf22-kube-api-access-48hqj\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.933862 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1516035e-4dd8-48ab-945a-09f7bee4bf22-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.933879 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.933894 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1516035e-4dd8-48ab-945a-09f7bee4bf22-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.973411 4741 generic.go:334] "Generic (PLEG): container finished" podID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerID="707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695" exitCode=0 Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.973437 4741 generic.go:334] "Generic (PLEG): container finished" podID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerID="9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7" exitCode=143 Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.973471 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1516035e-4dd8-48ab-945a-09f7bee4bf22","Type":"ContainerDied","Data":"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695"} Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.973498 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1516035e-4dd8-48ab-945a-09f7bee4bf22","Type":"ContainerDied","Data":"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7"} Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.973509 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1516035e-4dd8-48ab-945a-09f7bee4bf22","Type":"ContainerDied","Data":"6ca33e3a8b281d6e9780b0ee013dd9c8a83be986a2e3835d6ab9a4ddecfbd0cc"} Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.973522 4741 scope.go:117] "RemoveContainer" containerID="707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.973622 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.978987 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659f5d9f44-j8zgh" event={"ID":"58cfa8a3-5067-40ee-8e23-803a52d6b1b1","Type":"ContainerStarted","Data":"58644df2679ff5827ca80923ddee91fde9b4da2f9eb6e37ae982ca0c9572be55"} Sep 29 19:28:14 crc kubenswrapper[4741]: I0929 19:28:14.979043 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659f5d9f44-j8zgh" event={"ID":"58cfa8a3-5067-40ee-8e23-803a52d6b1b1","Type":"ContainerStarted","Data":"dab1bbdc6afa761fe9a849e8ae5ff2eb85c04e27d9ed56440a407f6e2e4c7bf5"} Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.011922 4741 scope.go:117] "RemoveContainer" containerID="9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.016910 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.030075 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.038347 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:15 crc kubenswrapper[4741]: E0929 19:28:15.038843 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.038864 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api" Sep 29 19:28:15 crc kubenswrapper[4741]: E0929 19:28:15.038885 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api-log" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.038892 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api-log" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.039063 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api-log" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.039082 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" containerName="cinder-api" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.040057 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.042911 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.043027 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.043088 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.048590 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.062755 4741 scope.go:117] "RemoveContainer" containerID="707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695" Sep 29 19:28:15 crc kubenswrapper[4741]: E0929 19:28:15.063429 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695\": container with ID starting with 707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695 not found: ID does not exist" containerID="707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.063550 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695"} err="failed to get container status \"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695\": rpc error: code = NotFound desc = could not find container \"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695\": container with ID starting with 707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695 not found: ID does not exist" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.063616 4741 scope.go:117] "RemoveContainer" containerID="9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7" Sep 29 19:28:15 crc kubenswrapper[4741]: E0929 19:28:15.064004 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7\": container with ID starting with 9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7 not found: ID does not exist" containerID="9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.064049 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7"} err="failed to get container status \"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7\": rpc error: code = NotFound desc = could not find container \"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7\": container with ID starting with 9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7 not found: ID does not exist" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.064076 4741 scope.go:117] "RemoveContainer" containerID="707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.064458 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695"} err="failed to get container status \"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695\": rpc error: code = NotFound desc = could not find container \"707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695\": container with ID starting with 707f181cd9a3695a0ffab1a7c5f665bae83e014c82c5d0e7e77f78a693fa7695 not found: ID does not exist" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.064510 4741 scope.go:117] "RemoveContainer" containerID="9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.064751 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7"} err="failed to get container status \"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7\": rpc error: code = NotFound desc = could not find container \"9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7\": container with ID starting with 9dde09f53aed31cd2df4f56517de61d98dc6a936803faad814fe5700a2361ae7 not found: ID does not exist" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.097621 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1516035e-4dd8-48ab-945a-09f7bee4bf22" path="/var/lib/kubelet/pods/1516035e-4dd8-48ab-945a-09f7bee4bf22/volumes" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.137527 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b65efd-d659-4b2d-9cee-0c57b08d029f-logs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.137831 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-scripts\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.137931 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.138008 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.138060 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3b65efd-d659-4b2d-9cee-0c57b08d029f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.138298 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.138338 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl6df\" (UniqueName: \"kubernetes.io/projected/e3b65efd-d659-4b2d-9cee-0c57b08d029f-kube-api-access-wl6df\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.138375 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.138427 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.241927 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b65efd-d659-4b2d-9cee-0c57b08d029f-logs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242103 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-scripts\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242146 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242189 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242215 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3b65efd-d659-4b2d-9cee-0c57b08d029f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242332 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242358 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b65efd-d659-4b2d-9cee-0c57b08d029f-logs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242411 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3b65efd-d659-4b2d-9cee-0c57b08d029f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242415 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl6df\" (UniqueName: \"kubernetes.io/projected/e3b65efd-d659-4b2d-9cee-0c57b08d029f-kube-api-access-wl6df\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242479 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.242501 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.246684 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-scripts\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.247146 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.247253 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.247874 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.248566 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.254696 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.265614 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl6df\" (UniqueName: \"kubernetes.io/projected/e3b65efd-d659-4b2d-9cee-0c57b08d029f-kube-api-access-wl6df\") pod \"cinder-api-0\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.371362 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.859625 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.992404 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3b65efd-d659-4b2d-9cee-0c57b08d029f","Type":"ContainerStarted","Data":"6999f3d4635471e5c45211f770e66bf963e5812a4d3ce22e0299954d1838c487"} Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.997763 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659f5d9f44-j8zgh" event={"ID":"58cfa8a3-5067-40ee-8e23-803a52d6b1b1","Type":"ContainerStarted","Data":"dfd2422ee682a56f33a264c928c2202071429b72837518805278e82828695484"} Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.997965 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:15 crc kubenswrapper[4741]: I0929 19:28:15.998003 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:16 crc kubenswrapper[4741]: I0929 19:28:16.026651 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-659f5d9f44-j8zgh" podStartSLOduration=4.026614531 podStartE2EDuration="4.026614531s" podCreationTimestamp="2025-09-29 19:28:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:16.0186646 +0000 UTC m=+1137.666453942" watchObservedRunningTime="2025-09-29 19:28:16.026614531 +0000 UTC m=+1137.674403863" Sep 29 19:28:17 crc kubenswrapper[4741]: I0929 19:28:17.024518 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3b65efd-d659-4b2d-9cee-0c57b08d029f","Type":"ContainerStarted","Data":"eac6f805571d33e47abcb8386b1e4a78317268ba124fe10a2e9376e94d5dd506"} Sep 29 19:28:17 crc kubenswrapper[4741]: I0929 19:28:17.095935 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.027828 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.043574 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3b65efd-d659-4b2d-9cee-0c57b08d029f","Type":"ContainerStarted","Data":"5747543db2d4686ff0ce2d66484ba05733793d16bc7f1a3c6bf312799d425845"} Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.043911 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.079599 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.079564544 podStartE2EDuration="3.079564544s" podCreationTimestamp="2025-09-29 19:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:18.067721309 +0000 UTC m=+1139.715510641" watchObservedRunningTime="2025-09-29 19:28:18.079564544 +0000 UTC m=+1139.727353886" Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.175746 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.465650 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.542248 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wtx87"] Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.542552 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" podUID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerName="dnsmasq-dns" containerID="cri-o://2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4" gracePeriod=10 Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.659937 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 29 19:28:18 crc kubenswrapper[4741]: I0929 19:28:18.702493 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.045014 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.057978 4741 generic.go:334] "Generic (PLEG): container finished" podID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerID="2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4" exitCode=0 Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.058053 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" event={"ID":"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32","Type":"ContainerDied","Data":"2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4"} Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.058116 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" event={"ID":"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32","Type":"ContainerDied","Data":"9c1ea77995031fcbca61b25ce56741ac0141e1383b317e43c946672661dba0c6"} Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.058135 4741 scope.go:117] "RemoveContainer" containerID="2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.058341 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wtx87" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.058902 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="probe" containerID="cri-o://e668bdb14dae689ea32169dbee26af2d538bda4ff099cd9c2757e3a67cdabfb8" gracePeriod=30 Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.058908 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="cinder-scheduler" containerID="cri-o://fc92aecb2dfb9850b26acbeeeaeffacdb85da5b76c5b3b7b8969e785ccce435e" gracePeriod=30 Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.086500 4741 scope.go:117] "RemoveContainer" containerID="d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.129922 4741 scope.go:117] "RemoveContainer" containerID="2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4" Sep 29 19:28:19 crc kubenswrapper[4741]: E0929 19:28:19.134706 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4\": container with ID starting with 2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4 not found: ID does not exist" containerID="2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.134759 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4"} err="failed to get container status \"2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4\": rpc error: code = NotFound desc = could not find container \"2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4\": container with ID starting with 2bd4c0bc595cfd78daca48a5287b752ed5b25bbe2bedbc5709ddf89f320470b4 not found: ID does not exist" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.134779 4741 scope.go:117] "RemoveContainer" containerID="d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200" Sep 29 19:28:19 crc kubenswrapper[4741]: E0929 19:28:19.136382 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200\": container with ID starting with d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200 not found: ID does not exist" containerID="d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.136423 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200"} err="failed to get container status \"d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200\": rpc error: code = NotFound desc = could not find container \"d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200\": container with ID starting with d5ca5a69c0fcd03c7247f132957f3f6ffc1e5b767464b83c679a8bcf1ca6d200 not found: ID does not exist" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.172698 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.214653 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-swift-storage-0\") pod \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.214716 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q52mq\" (UniqueName: \"kubernetes.io/projected/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-kube-api-access-q52mq\") pod \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.214741 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-sb\") pod \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.214769 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-nb\") pod \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.214825 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-config\") pod \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.214919 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-svc\") pod \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\" (UID: \"2de732e3-97b2-49b4-9ea2-5a48d2a6cb32\") " Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.228736 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-kube-api-access-q52mq" (OuterVolumeSpecName: "kube-api-access-q52mq") pod "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" (UID: "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32"). InnerVolumeSpecName "kube-api-access-q52mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.251751 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86bbf6b688-lcffw"] Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.252146 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-86bbf6b688-lcffw" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-api" containerID="cri-o://20400580a7b1bb2a730ab3cf2404a7dde79dceaf6d945e1e6553f0c844c4aa95" gracePeriod=30 Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.252501 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-86bbf6b688-lcffw" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-httpd" containerID="cri-o://e0350f3b0dc57e12e5f350135bdb6938f5c4bd6180fecfcf91eb04371e1d439b" gracePeriod=30 Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.306035 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-config" (OuterVolumeSpecName: "config") pod "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" (UID: "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.309019 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" (UID: "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.309109 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" (UID: "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.317037 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.317059 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q52mq\" (UniqueName: \"kubernetes.io/projected/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-kube-api-access-q52mq\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.317071 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.317080 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.321930 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" (UID: "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.330123 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" (UID: "2de732e3-97b2-49b4-9ea2-5a48d2a6cb32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.409431 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wtx87"] Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.415931 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wtx87"] Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.418849 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:19 crc kubenswrapper[4741]: I0929 19:28:19.418881 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:20 crc kubenswrapper[4741]: I0929 19:28:20.072822 4741 generic.go:334] "Generic (PLEG): container finished" podID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerID="e668bdb14dae689ea32169dbee26af2d538bda4ff099cd9c2757e3a67cdabfb8" exitCode=0 Sep 29 19:28:20 crc kubenswrapper[4741]: I0929 19:28:20.073057 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8b970fb5-e6f0-44e0-9dd1-6391661533ed","Type":"ContainerDied","Data":"e668bdb14dae689ea32169dbee26af2d538bda4ff099cd9c2757e3a67cdabfb8"} Sep 29 19:28:20 crc kubenswrapper[4741]: I0929 19:28:20.078193 4741 generic.go:334] "Generic (PLEG): container finished" podID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerID="e0350f3b0dc57e12e5f350135bdb6938f5c4bd6180fecfcf91eb04371e1d439b" exitCode=0 Sep 29 19:28:20 crc kubenswrapper[4741]: I0929 19:28:20.078226 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86bbf6b688-lcffw" event={"ID":"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1","Type":"ContainerDied","Data":"e0350f3b0dc57e12e5f350135bdb6938f5c4bd6180fecfcf91eb04371e1d439b"} Sep 29 19:28:20 crc kubenswrapper[4741]: I0929 19:28:20.850264 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:28:20 crc kubenswrapper[4741]: I0929 19:28:20.871053 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:28:20 crc kubenswrapper[4741]: I0929 19:28:20.895144 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:21 crc kubenswrapper[4741]: I0929 19:28:21.100411 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" path="/var/lib/kubelet/pods/2de732e3-97b2-49b4-9ea2-5a48d2a6cb32/volumes" Sep 29 19:28:22 crc kubenswrapper[4741]: I0929 19:28:22.354379 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:28:22 crc kubenswrapper[4741]: I0929 19:28:22.442398 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5c8494f4cd-9bsk9"] Sep 29 19:28:22 crc kubenswrapper[4741]: I0929 19:28:22.442618 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5c8494f4cd-9bsk9" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api-log" containerID="cri-o://7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8" gracePeriod=30 Sep 29 19:28:22 crc kubenswrapper[4741]: I0929 19:28:22.443000 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5c8494f4cd-9bsk9" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api" containerID="cri-o://75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829" gracePeriod=30 Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.121664 4741 generic.go:334] "Generic (PLEG): container finished" podID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerID="7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8" exitCode=143 Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.121994 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c8494f4cd-9bsk9" event={"ID":"e918ae11-61d6-4766-b731-dc3b9f9c6abc","Type":"ContainerDied","Data":"7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8"} Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.123460 4741 generic.go:334] "Generic (PLEG): container finished" podID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerID="fc92aecb2dfb9850b26acbeeeaeffacdb85da5b76c5b3b7b8969e785ccce435e" exitCode=0 Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.123476 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8b970fb5-e6f0-44e0-9dd1-6391661533ed","Type":"ContainerDied","Data":"fc92aecb2dfb9850b26acbeeeaeffacdb85da5b76c5b3b7b8969e785ccce435e"} Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.308651 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.411521 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data\") pod \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.411616 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b970fb5-e6f0-44e0-9dd1-6391661533ed-etc-machine-id\") pod \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.411679 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-combined-ca-bundle\") pod \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.411754 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qhz7\" (UniqueName: \"kubernetes.io/projected/8b970fb5-e6f0-44e0-9dd1-6391661533ed-kube-api-access-2qhz7\") pod \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.411854 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-scripts\") pod \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.411953 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data-custom\") pod \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\" (UID: \"8b970fb5-e6f0-44e0-9dd1-6391661533ed\") " Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.413030 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b970fb5-e6f0-44e0-9dd1-6391661533ed-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8b970fb5-e6f0-44e0-9dd1-6391661533ed" (UID: "8b970fb5-e6f0-44e0-9dd1-6391661533ed"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.421933 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b970fb5-e6f0-44e0-9dd1-6391661533ed-kube-api-access-2qhz7" (OuterVolumeSpecName: "kube-api-access-2qhz7") pod "8b970fb5-e6f0-44e0-9dd1-6391661533ed" (UID: "8b970fb5-e6f0-44e0-9dd1-6391661533ed"). InnerVolumeSpecName "kube-api-access-2qhz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.426640 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8b970fb5-e6f0-44e0-9dd1-6391661533ed" (UID: "8b970fb5-e6f0-44e0-9dd1-6391661533ed"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.428471 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-scripts" (OuterVolumeSpecName: "scripts") pod "8b970fb5-e6f0-44e0-9dd1-6391661533ed" (UID: "8b970fb5-e6f0-44e0-9dd1-6391661533ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.466110 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b970fb5-e6f0-44e0-9dd1-6391661533ed" (UID: "8b970fb5-e6f0-44e0-9dd1-6391661533ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.503260 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data" (OuterVolumeSpecName: "config-data") pod "8b970fb5-e6f0-44e0-9dd1-6391661533ed" (UID: "8b970fb5-e6f0-44e0-9dd1-6391661533ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.514372 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.514408 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qhz7\" (UniqueName: \"kubernetes.io/projected/8b970fb5-e6f0-44e0-9dd1-6391661533ed-kube-api-access-2qhz7\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.514422 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.514430 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.514438 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b970fb5-e6f0-44e0-9dd1-6391661533ed-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:23 crc kubenswrapper[4741]: I0929 19:28:23.514446 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b970fb5-e6f0-44e0-9dd1-6391661533ed-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.134014 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.134028 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8b970fb5-e6f0-44e0-9dd1-6391661533ed","Type":"ContainerDied","Data":"32e0b9115f10f276b97f9ba483d6e0b1f197463f2f2757a4caad06a4e9883eff"} Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.134422 4741 scope.go:117] "RemoveContainer" containerID="e668bdb14dae689ea32169dbee26af2d538bda4ff099cd9c2757e3a67cdabfb8" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.138725 4741 generic.go:334] "Generic (PLEG): container finished" podID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerID="20400580a7b1bb2a730ab3cf2404a7dde79dceaf6d945e1e6553f0c844c4aa95" exitCode=0 Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.138759 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86bbf6b688-lcffw" event={"ID":"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1","Type":"ContainerDied","Data":"20400580a7b1bb2a730ab3cf2404a7dde79dceaf6d945e1e6553f0c844c4aa95"} Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.180703 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.185012 4741 scope.go:117] "RemoveContainer" containerID="fc92aecb2dfb9850b26acbeeeaeffacdb85da5b76c5b3b7b8969e785ccce435e" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.200821 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.211986 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:24 crc kubenswrapper[4741]: E0929 19:28:24.212412 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="probe" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.212430 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="probe" Sep 29 19:28:24 crc kubenswrapper[4741]: E0929 19:28:24.212445 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerName="init" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.212454 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerName="init" Sep 29 19:28:24 crc kubenswrapper[4741]: E0929 19:28:24.212483 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="cinder-scheduler" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.212490 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="cinder-scheduler" Sep 29 19:28:24 crc kubenswrapper[4741]: E0929 19:28:24.212506 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerName="dnsmasq-dns" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.212514 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerName="dnsmasq-dns" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.212732 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="probe" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.212751 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" containerName="cinder-scheduler" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.212768 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de732e3-97b2-49b4-9ea2-5a48d2a6cb32" containerName="dnsmasq-dns" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.213721 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.216220 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.228983 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c907b3ee-67ab-4d29-a02d-96137c124ff4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.229079 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.229148 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-scripts\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.229194 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.229238 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dqxx\" (UniqueName: \"kubernetes.io/projected/c907b3ee-67ab-4d29-a02d-96137c124ff4-kube-api-access-2dqxx\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.229287 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.235440 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.333415 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c907b3ee-67ab-4d29-a02d-96137c124ff4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.333496 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.333538 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-scripts\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.333548 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c907b3ee-67ab-4d29-a02d-96137c124ff4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.333572 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.333700 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dqxx\" (UniqueName: \"kubernetes.io/projected/c907b3ee-67ab-4d29-a02d-96137c124ff4-kube-api-access-2dqxx\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.333825 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.340981 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.343698 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.348437 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.348709 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-scripts\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.354808 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dqxx\" (UniqueName: \"kubernetes.io/projected/c907b3ee-67ab-4d29-a02d-96137c124ff4-kube-api-access-2dqxx\") pod \"cinder-scheduler-0\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.545672 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.725815 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.846083 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnqtv\" (UniqueName: \"kubernetes.io/projected/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-kube-api-access-tnqtv\") pod \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.846233 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-httpd-config\") pod \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.846312 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-combined-ca-bundle\") pod \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.846380 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-ovndb-tls-certs\") pod \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.846422 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-config\") pod \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\" (UID: \"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1\") " Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.851645 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" (UID: "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.857625 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-kube-api-access-tnqtv" (OuterVolumeSpecName: "kube-api-access-tnqtv") pod "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" (UID: "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1"). InnerVolumeSpecName "kube-api-access-tnqtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.947468 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.947494 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnqtv\" (UniqueName: \"kubernetes.io/projected/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-kube-api-access-tnqtv\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.951240 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" (UID: "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:24 crc kubenswrapper[4741]: I0929 19:28:24.958622 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-config" (OuterVolumeSpecName: "config") pod "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" (UID: "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.020770 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" (UID: "dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.048369 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.048421 4741 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.048431 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.095034 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b970fb5-e6f0-44e0-9dd1-6391661533ed" path="/var/lib/kubelet/pods/8b970fb5-e6f0-44e0-9dd1-6391661533ed/volumes" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.096172 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.153209 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c907b3ee-67ab-4d29-a02d-96137c124ff4","Type":"ContainerStarted","Data":"233545f1b736491909602c4210ef4aa4ef9018b304a5faf889117f508dafcc56"} Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.164709 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86bbf6b688-lcffw" event={"ID":"dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1","Type":"ContainerDied","Data":"b7426772344e481587347259225cdde7277e7a3d2a9b97397bc8febfb8493777"} Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.164773 4741 scope.go:117] "RemoveContainer" containerID="e0350f3b0dc57e12e5f350135bdb6938f5c4bd6180fecfcf91eb04371e1d439b" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.165041 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86bbf6b688-lcffw" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.206487 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86bbf6b688-lcffw"] Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.218272 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-86bbf6b688-lcffw"] Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.220111 4741 scope.go:117] "RemoveContainer" containerID="20400580a7b1bb2a730ab3cf2404a7dde79dceaf6d945e1e6553f0c844c4aa95" Sep 29 19:28:25 crc kubenswrapper[4741]: I0929 19:28:25.645984 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.077351 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.196022 4741 generic.go:334] "Generic (PLEG): container finished" podID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerID="75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829" exitCode=0 Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.196165 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c8494f4cd-9bsk9" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.196121 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c8494f4cd-9bsk9" event={"ID":"e918ae11-61d6-4766-b731-dc3b9f9c6abc","Type":"ContainerDied","Data":"75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829"} Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.196231 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c8494f4cd-9bsk9" event={"ID":"e918ae11-61d6-4766-b731-dc3b9f9c6abc","Type":"ContainerDied","Data":"20670d470185d459ac22f7e767252d4aaa388b331ed1697f0c54ddd48cffae85"} Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.196252 4741 scope.go:117] "RemoveContainer" containerID="75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.200793 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c907b3ee-67ab-4d29-a02d-96137c124ff4","Type":"ContainerStarted","Data":"a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04"} Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.253337 4741 scope.go:117] "RemoveContainer" containerID="7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.269799 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e918ae11-61d6-4766-b731-dc3b9f9c6abc-logs\") pod \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.269947 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pvwg\" (UniqueName: \"kubernetes.io/projected/e918ae11-61d6-4766-b731-dc3b9f9c6abc-kube-api-access-7pvwg\") pod \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.269970 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-combined-ca-bundle\") pod \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.269987 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data\") pod \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.270045 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data-custom\") pod \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\" (UID: \"e918ae11-61d6-4766-b731-dc3b9f9c6abc\") " Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.270462 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e918ae11-61d6-4766-b731-dc3b9f9c6abc-logs" (OuterVolumeSpecName: "logs") pod "e918ae11-61d6-4766-b731-dc3b9f9c6abc" (UID: "e918ae11-61d6-4766-b731-dc3b9f9c6abc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.276636 4741 scope.go:117] "RemoveContainer" containerID="75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.277482 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e918ae11-61d6-4766-b731-dc3b9f9c6abc" (UID: "e918ae11-61d6-4766-b731-dc3b9f9c6abc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.277558 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e918ae11-61d6-4766-b731-dc3b9f9c6abc-kube-api-access-7pvwg" (OuterVolumeSpecName: "kube-api-access-7pvwg") pod "e918ae11-61d6-4766-b731-dc3b9f9c6abc" (UID: "e918ae11-61d6-4766-b731-dc3b9f9c6abc"). InnerVolumeSpecName "kube-api-access-7pvwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:26 crc kubenswrapper[4741]: E0929 19:28:26.277549 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829\": container with ID starting with 75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829 not found: ID does not exist" containerID="75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.277598 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829"} err="failed to get container status \"75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829\": rpc error: code = NotFound desc = could not find container \"75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829\": container with ID starting with 75d611a330bd79269155e5cdfd1ab3a04887d0ec8ac9cad95653f50d3ceb2829 not found: ID does not exist" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.277621 4741 scope.go:117] "RemoveContainer" containerID="7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8" Sep 29 19:28:26 crc kubenswrapper[4741]: E0929 19:28:26.277905 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8\": container with ID starting with 7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8 not found: ID does not exist" containerID="7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.277930 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8"} err="failed to get container status \"7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8\": rpc error: code = NotFound desc = could not find container \"7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8\": container with ID starting with 7ab20c715fd271236d5b1df077ae9e32a534a2ff30328dc1eaada52a16c291f8 not found: ID does not exist" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.297025 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e918ae11-61d6-4766-b731-dc3b9f9c6abc" (UID: "e918ae11-61d6-4766-b731-dc3b9f9c6abc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.328927 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data" (OuterVolumeSpecName: "config-data") pod "e918ae11-61d6-4766-b731-dc3b9f9c6abc" (UID: "e918ae11-61d6-4766-b731-dc3b9f9c6abc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.372332 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e918ae11-61d6-4766-b731-dc3b9f9c6abc-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.372367 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pvwg\" (UniqueName: \"kubernetes.io/projected/e918ae11-61d6-4766-b731-dc3b9f9c6abc-kube-api-access-7pvwg\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.372382 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.372405 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.372417 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e918ae11-61d6-4766-b731-dc3b9f9c6abc-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.566473 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5c8494f4cd-9bsk9"] Sep 29 19:28:26 crc kubenswrapper[4741]: I0929 19:28:26.572922 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5c8494f4cd-9bsk9"] Sep 29 19:28:27 crc kubenswrapper[4741]: I0929 19:28:27.095879 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" path="/var/lib/kubelet/pods/dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1/volumes" Sep 29 19:28:27 crc kubenswrapper[4741]: I0929 19:28:27.096737 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" path="/var/lib/kubelet/pods/e918ae11-61d6-4766-b731-dc3b9f9c6abc/volumes" Sep 29 19:28:27 crc kubenswrapper[4741]: I0929 19:28:27.214580 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c907b3ee-67ab-4d29-a02d-96137c124ff4","Type":"ContainerStarted","Data":"5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7"} Sep 29 19:28:27 crc kubenswrapper[4741]: I0929 19:28:27.239904 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.239885475 podStartE2EDuration="3.239885475s" podCreationTimestamp="2025-09-29 19:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:27.23246533 +0000 UTC m=+1148.880254662" watchObservedRunningTime="2025-09-29 19:28:27.239885475 +0000 UTC m=+1148.887674807" Sep 29 19:28:27 crc kubenswrapper[4741]: I0929 19:28:27.516036 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.144381 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 29 19:28:29 crc kubenswrapper[4741]: E0929 19:28:29.145961 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.146031 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api" Sep 29 19:28:29 crc kubenswrapper[4741]: E0929 19:28:29.146091 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-httpd" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.146140 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-httpd" Sep 29 19:28:29 crc kubenswrapper[4741]: E0929 19:28:29.146193 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-api" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.146338 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-api" Sep 29 19:28:29 crc kubenswrapper[4741]: E0929 19:28:29.146435 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api-log" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.146501 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api-log" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.146900 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.146992 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-api" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.147080 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="dccc1fd0-0b9f-41a0-89fe-8dcae0f281c1" containerName="neutron-httpd" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.147150 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e918ae11-61d6-4766-b731-dc3b9f9c6abc" containerName="barbican-api-log" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.147866 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.156031 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.156950 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.158594 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-n7q66" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.167193 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.222803 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p9s5\" (UniqueName: \"kubernetes.io/projected/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-kube-api-access-7p9s5\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.222852 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.222887 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.223002 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config-secret\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.324981 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.325192 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config-secret\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.325353 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p9s5\" (UniqueName: \"kubernetes.io/projected/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-kube-api-access-7p9s5\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.325470 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.325858 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.330455 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.332605 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config-secret\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.345458 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p9s5\" (UniqueName: \"kubernetes.io/projected/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-kube-api-access-7p9s5\") pod \"openstackclient\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.480024 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.546676 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 29 19:28:29 crc kubenswrapper[4741]: I0929 19:28:29.947123 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 19:28:29 crc kubenswrapper[4741]: W0929 19:28:29.949146 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc0eaaf0_97ff_47ea_a15d_0024fbc96439.slice/crio-e164a12db5e178895df9a4bd2225b39000dc1732120109429425f44590f7536a WatchSource:0}: Error finding container e164a12db5e178895df9a4bd2225b39000dc1732120109429425f44590f7536a: Status 404 returned error can't find the container with id e164a12db5e178895df9a4bd2225b39000dc1732120109429425f44590f7536a Sep 29 19:28:30 crc kubenswrapper[4741]: I0929 19:28:30.264517 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"fc0eaaf0-97ff-47ea-a15d-0024fbc96439","Type":"ContainerStarted","Data":"e164a12db5e178895df9a4bd2225b39000dc1732120109429425f44590f7536a"} Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.608343 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.608873 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-central-agent" containerID="cri-o://d04ab1347974022043dd25c8315eebd22bbe9958a62469dfe66ccd082f29ace1" gracePeriod=30 Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.608987 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-notification-agent" containerID="cri-o://698002fd90572118b08152d0ec223b9498eef909ee7b1674d236a8b8b8fc3f79" gracePeriod=30 Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.608984 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="proxy-httpd" containerID="cri-o://03c697990d9bf6337f6a018465b8ba174c224f446de542677af9f64c31f0c004" gracePeriod=30 Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.609016 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="sg-core" containerID="cri-o://1c0079aa744f1248402f2e7a77e71d07d8e514fd609bf0af6d642f6a7997bb69" gracePeriod=30 Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.617929 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.153:3000/\": EOF" Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.738760 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:28:31 crc kubenswrapper[4741]: I0929 19:28:31.738822 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.287738 4741 generic.go:334] "Generic (PLEG): container finished" podID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerID="03c697990d9bf6337f6a018465b8ba174c224f446de542677af9f64c31f0c004" exitCode=0 Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.288024 4741 generic.go:334] "Generic (PLEG): container finished" podID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerID="1c0079aa744f1248402f2e7a77e71d07d8e514fd609bf0af6d642f6a7997bb69" exitCode=2 Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.288034 4741 generic.go:334] "Generic (PLEG): container finished" podID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerID="d04ab1347974022043dd25c8315eebd22bbe9958a62469dfe66ccd082f29ace1" exitCode=0 Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.287817 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerDied","Data":"03c697990d9bf6337f6a018465b8ba174c224f446de542677af9f64c31f0c004"} Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.288069 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerDied","Data":"1c0079aa744f1248402f2e7a77e71d07d8e514fd609bf0af6d642f6a7997bb69"} Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.288082 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerDied","Data":"d04ab1347974022043dd25c8315eebd22bbe9958a62469dfe66ccd082f29ace1"} Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.955725 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-74d7c6f55-9trnx"] Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.957704 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.960966 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.961160 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.961289 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Sep 29 19:28:32 crc kubenswrapper[4741]: I0929 19:28:32.968132 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-74d7c6f55-9trnx"] Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.085868 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-combined-ca-bundle\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.085956 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-config-data\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.086202 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-internal-tls-certs\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.086236 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q8fl\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-kube-api-access-8q8fl\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.086318 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-run-httpd\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.086409 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-etc-swift\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.086463 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-public-tls-certs\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.086522 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-log-httpd\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.188780 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-etc-swift\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.188901 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-public-tls-certs\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.188959 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-log-httpd\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.189063 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-combined-ca-bundle\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.189096 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-config-data\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.189197 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-internal-tls-certs\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.189232 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q8fl\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-kube-api-access-8q8fl\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.189332 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-run-httpd\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.190292 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-run-httpd\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.191165 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-log-httpd\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.195977 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-public-tls-certs\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.196030 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-etc-swift\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.198667 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-combined-ca-bundle\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.201266 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-config-data\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.204525 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-internal-tls-certs\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.211754 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q8fl\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-kube-api-access-8q8fl\") pod \"swift-proxy-74d7c6f55-9trnx\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:33 crc kubenswrapper[4741]: I0929 19:28:33.292277 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:34 crc kubenswrapper[4741]: I0929 19:28:34.746960 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 29 19:28:36 crc kubenswrapper[4741]: I0929 19:28:36.206841 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.153:3000/\": dial tcp 10.217.0.153:3000: connect: connection refused" Sep 29 19:28:36 crc kubenswrapper[4741]: I0929 19:28:36.337472 4741 generic.go:334] "Generic (PLEG): container finished" podID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerID="698002fd90572118b08152d0ec223b9498eef909ee7b1674d236a8b8b8fc3f79" exitCode=0 Sep 29 19:28:36 crc kubenswrapper[4741]: I0929 19:28:36.337520 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerDied","Data":"698002fd90572118b08152d0ec223b9498eef909ee7b1674d236a8b8b8fc3f79"} Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.503867 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5ffsc"] Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.510919 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5ffsc" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.545980 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5ffsc"] Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.600552 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-2nvjd"] Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.601712 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2nvjd" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.607047 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2nvjd"] Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.676971 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.695205 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-gcx8j"] Sep 29 19:28:38 crc kubenswrapper[4741]: E0929 19:28:38.696070 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-central-agent" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696094 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-central-agent" Sep 29 19:28:38 crc kubenswrapper[4741]: E0929 19:28:38.696116 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="proxy-httpd" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696153 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="proxy-httpd" Sep 29 19:28:38 crc kubenswrapper[4741]: E0929 19:28:38.696171 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="sg-core" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696181 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="sg-core" Sep 29 19:28:38 crc kubenswrapper[4741]: E0929 19:28:38.696240 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-notification-agent" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696251 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-notification-agent" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696678 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="sg-core" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696718 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="proxy-httpd" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696769 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-notification-agent" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.696790 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" containerName="ceilometer-central-agent" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.700011 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gcx8j" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.702891 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gcx8j"] Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.715925 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pn5c\" (UniqueName: \"kubernetes.io/projected/8f495781-1870-461e-b12d-ddf347b0a73c-kube-api-access-9pn5c\") pod \"nova-api-db-create-5ffsc\" (UID: \"8f495781-1870-461e-b12d-ddf347b0a73c\") " pod="openstack/nova-api-db-create-5ffsc" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.716008 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghzr2\" (UniqueName: \"kubernetes.io/projected/4f1730aa-2bc9-4005-88fb-4634745de0aa-kube-api-access-ghzr2\") pod \"nova-cell0-db-create-2nvjd\" (UID: \"4f1730aa-2bc9-4005-88fb-4634745de0aa\") " pod="openstack/nova-cell0-db-create-2nvjd" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.816813 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-log-httpd\") pod \"3390cf5f-a926-4ea3-9fae-d560cd99789a\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.816896 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-sg-core-conf-yaml\") pod \"3390cf5f-a926-4ea3-9fae-d560cd99789a\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.816972 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-scripts\") pod \"3390cf5f-a926-4ea3-9fae-d560cd99789a\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817000 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-config-data\") pod \"3390cf5f-a926-4ea3-9fae-d560cd99789a\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817205 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2wpx\" (UniqueName: \"kubernetes.io/projected/3390cf5f-a926-4ea3-9fae-d560cd99789a-kube-api-access-d2wpx\") pod \"3390cf5f-a926-4ea3-9fae-d560cd99789a\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817260 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-run-httpd\") pod \"3390cf5f-a926-4ea3-9fae-d560cd99789a\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817331 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-combined-ca-bundle\") pod \"3390cf5f-a926-4ea3-9fae-d560cd99789a\" (UID: \"3390cf5f-a926-4ea3-9fae-d560cd99789a\") " Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817423 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3390cf5f-a926-4ea3-9fae-d560cd99789a" (UID: "3390cf5f-a926-4ea3-9fae-d560cd99789a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817591 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3390cf5f-a926-4ea3-9fae-d560cd99789a" (UID: "3390cf5f-a926-4ea3-9fae-d560cd99789a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817892 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjxjp\" (UniqueName: \"kubernetes.io/projected/ce0c5a54-c02f-4f97-a0ea-54383f8b792c-kube-api-access-zjxjp\") pod \"nova-cell1-db-create-gcx8j\" (UID: \"ce0c5a54-c02f-4f97-a0ea-54383f8b792c\") " pod="openstack/nova-cell1-db-create-gcx8j" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.817972 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pn5c\" (UniqueName: \"kubernetes.io/projected/8f495781-1870-461e-b12d-ddf347b0a73c-kube-api-access-9pn5c\") pod \"nova-api-db-create-5ffsc\" (UID: \"8f495781-1870-461e-b12d-ddf347b0a73c\") " pod="openstack/nova-api-db-create-5ffsc" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.818012 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghzr2\" (UniqueName: \"kubernetes.io/projected/4f1730aa-2bc9-4005-88fb-4634745de0aa-kube-api-access-ghzr2\") pod \"nova-cell0-db-create-2nvjd\" (UID: \"4f1730aa-2bc9-4005-88fb-4634745de0aa\") " pod="openstack/nova-cell0-db-create-2nvjd" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.818168 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.818181 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3390cf5f-a926-4ea3-9fae-d560cd99789a-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.821971 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-scripts" (OuterVolumeSpecName: "scripts") pod "3390cf5f-a926-4ea3-9fae-d560cd99789a" (UID: "3390cf5f-a926-4ea3-9fae-d560cd99789a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.822794 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3390cf5f-a926-4ea3-9fae-d560cd99789a-kube-api-access-d2wpx" (OuterVolumeSpecName: "kube-api-access-d2wpx") pod "3390cf5f-a926-4ea3-9fae-d560cd99789a" (UID: "3390cf5f-a926-4ea3-9fae-d560cd99789a"). InnerVolumeSpecName "kube-api-access-d2wpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.833785 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pn5c\" (UniqueName: \"kubernetes.io/projected/8f495781-1870-461e-b12d-ddf347b0a73c-kube-api-access-9pn5c\") pod \"nova-api-db-create-5ffsc\" (UID: \"8f495781-1870-461e-b12d-ddf347b0a73c\") " pod="openstack/nova-api-db-create-5ffsc" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.834224 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghzr2\" (UniqueName: \"kubernetes.io/projected/4f1730aa-2bc9-4005-88fb-4634745de0aa-kube-api-access-ghzr2\") pod \"nova-cell0-db-create-2nvjd\" (UID: \"4f1730aa-2bc9-4005-88fb-4634745de0aa\") " pod="openstack/nova-cell0-db-create-2nvjd" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.841282 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3390cf5f-a926-4ea3-9fae-d560cd99789a" (UID: "3390cf5f-a926-4ea3-9fae-d560cd99789a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.843954 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5ffsc" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.920283 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjxjp\" (UniqueName: \"kubernetes.io/projected/ce0c5a54-c02f-4f97-a0ea-54383f8b792c-kube-api-access-zjxjp\") pod \"nova-cell1-db-create-gcx8j\" (UID: \"ce0c5a54-c02f-4f97-a0ea-54383f8b792c\") " pod="openstack/nova-cell1-db-create-gcx8j" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.920405 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.920417 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.920425 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2wpx\" (UniqueName: \"kubernetes.io/projected/3390cf5f-a926-4ea3-9fae-d560cd99789a-kube-api-access-d2wpx\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.930241 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2nvjd" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.935061 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-config-data" (OuterVolumeSpecName: "config-data") pod "3390cf5f-a926-4ea3-9fae-d560cd99789a" (UID: "3390cf5f-a926-4ea3-9fae-d560cd99789a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.937945 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjxjp\" (UniqueName: \"kubernetes.io/projected/ce0c5a54-c02f-4f97-a0ea-54383f8b792c-kube-api-access-zjxjp\") pod \"nova-cell1-db-create-gcx8j\" (UID: \"ce0c5a54-c02f-4f97-a0ea-54383f8b792c\") " pod="openstack/nova-cell1-db-create-gcx8j" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.949159 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3390cf5f-a926-4ea3-9fae-d560cd99789a" (UID: "3390cf5f-a926-4ea3-9fae-d560cd99789a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:38 crc kubenswrapper[4741]: I0929 19:28:38.986760 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-74d7c6f55-9trnx"] Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.022672 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.022697 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3390cf5f-a926-4ea3-9fae-d560cd99789a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.029375 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gcx8j" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.375113 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.375115 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3390cf5f-a926-4ea3-9fae-d560cd99789a","Type":"ContainerDied","Data":"66f4828960adaec5a0e0c33c00725b6f6c639909f811bd688f25a56104d89fce"} Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.375620 4741 scope.go:117] "RemoveContainer" containerID="03c697990d9bf6337f6a018465b8ba174c224f446de542677af9f64c31f0c004" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.377980 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d7c6f55-9trnx" event={"ID":"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe","Type":"ContainerStarted","Data":"cce06dd4a20cab8469d6440e5d832029b66e459b52c5ff9b817bb86dbe5b5145"} Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.378016 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d7c6f55-9trnx" event={"ID":"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe","Type":"ContainerStarted","Data":"7c8e4008fd8e4be8cdfa645117d0791e6517e3b394e16325326266514ea489ee"} Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.382447 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"fc0eaaf0-97ff-47ea-a15d-0024fbc96439","Type":"ContainerStarted","Data":"a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead"} Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.400413 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5ffsc"] Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.402003 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.988602733 podStartE2EDuration="10.401984364s" podCreationTimestamp="2025-09-29 19:28:29 +0000 UTC" firstStartedPulling="2025-09-29 19:28:29.954551503 +0000 UTC m=+1151.602340835" lastFinishedPulling="2025-09-29 19:28:38.367933134 +0000 UTC m=+1160.015722466" observedRunningTime="2025-09-29 19:28:39.400002871 +0000 UTC m=+1161.047792203" watchObservedRunningTime="2025-09-29 19:28:39.401984364 +0000 UTC m=+1161.049773706" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.452876 4741 scope.go:117] "RemoveContainer" containerID="1c0079aa744f1248402f2e7a77e71d07d8e514fd609bf0af6d642f6a7997bb69" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.503162 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2nvjd"] Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.515949 4741 scope.go:117] "RemoveContainer" containerID="698002fd90572118b08152d0ec223b9498eef909ee7b1674d236a8b8b8fc3f79" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.520434 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:39 crc kubenswrapper[4741]: W0929 19:28:39.523254 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f1730aa_2bc9_4005_88fb_4634745de0aa.slice/crio-ee1810932b65cb9ea5ea50cf322354a512b744e2515658f29f60ac86850c83c9 WatchSource:0}: Error finding container ee1810932b65cb9ea5ea50cf322354a512b744e2515658f29f60ac86850c83c9: Status 404 returned error can't find the container with id ee1810932b65cb9ea5ea50cf322354a512b744e2515658f29f60ac86850c83c9 Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.536817 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.557346 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.560997 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.568807 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gcx8j"] Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.577902 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.579125 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.579370 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.647287 4741 scope.go:117] "RemoveContainer" containerID="d04ab1347974022043dd25c8315eebd22bbe9958a62469dfe66ccd082f29ace1" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.739766 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.739927 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-config-data\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.740025 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-log-httpd\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.740084 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-scripts\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.740114 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-run-httpd\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.740174 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.740831 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s769m\" (UniqueName: \"kubernetes.io/projected/864c8ce0-8c53-4d62-88fd-635c425dd5d6-kube-api-access-s769m\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842197 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s769m\" (UniqueName: \"kubernetes.io/projected/864c8ce0-8c53-4d62-88fd-635c425dd5d6-kube-api-access-s769m\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842341 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842431 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-config-data\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842502 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-log-httpd\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842554 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-scripts\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842589 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-run-httpd\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.842987 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-log-httpd\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.843052 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-run-httpd\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.847329 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-config-data\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.847609 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.847819 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.851721 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-scripts\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.867204 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s769m\" (UniqueName: \"kubernetes.io/projected/864c8ce0-8c53-4d62-88fd-635c425dd5d6-kube-api-access-s769m\") pod \"ceilometer-0\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " pod="openstack/ceilometer-0" Sep 29 19:28:39 crc kubenswrapper[4741]: I0929 19:28:39.912104 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.370286 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:40 crc kubenswrapper[4741]: W0929 19:28:40.375493 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod864c8ce0_8c53_4d62_88fd_635c425dd5d6.slice/crio-54c8f7a3cd287419b7da5b9edf1fb0a25514af86b30666d927b5ecf11dbc4b4c WatchSource:0}: Error finding container 54c8f7a3cd287419b7da5b9edf1fb0a25514af86b30666d927b5ecf11dbc4b4c: Status 404 returned error can't find the container with id 54c8f7a3cd287419b7da5b9edf1fb0a25514af86b30666d927b5ecf11dbc4b4c Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.392076 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d7c6f55-9trnx" event={"ID":"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe","Type":"ContainerStarted","Data":"5488046b5c71d7ada89c02641c087812ffa5561b258e4750ab91f6cd07383d28"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.392167 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.392221 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.394913 4741 generic.go:334] "Generic (PLEG): container finished" podID="4f1730aa-2bc9-4005-88fb-4634745de0aa" containerID="5adf7bb9a660c6c77e2e4c1e061e23a7547defbfb032ae8b851205e03cab997b" exitCode=0 Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.394959 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2nvjd" event={"ID":"4f1730aa-2bc9-4005-88fb-4634745de0aa","Type":"ContainerDied","Data":"5adf7bb9a660c6c77e2e4c1e061e23a7547defbfb032ae8b851205e03cab997b"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.394986 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2nvjd" event={"ID":"4f1730aa-2bc9-4005-88fb-4634745de0aa","Type":"ContainerStarted","Data":"ee1810932b65cb9ea5ea50cf322354a512b744e2515658f29f60ac86850c83c9"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.396987 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerStarted","Data":"54c8f7a3cd287419b7da5b9edf1fb0a25514af86b30666d927b5ecf11dbc4b4c"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.398403 4741 generic.go:334] "Generic (PLEG): container finished" podID="ce0c5a54-c02f-4f97-a0ea-54383f8b792c" containerID="73c176af015738a2be3c557c9aba6b6eb0d0e9b25133c488196d8f50acaa001c" exitCode=0 Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.398451 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gcx8j" event={"ID":"ce0c5a54-c02f-4f97-a0ea-54383f8b792c","Type":"ContainerDied","Data":"73c176af015738a2be3c557c9aba6b6eb0d0e9b25133c488196d8f50acaa001c"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.398468 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gcx8j" event={"ID":"ce0c5a54-c02f-4f97-a0ea-54383f8b792c","Type":"ContainerStarted","Data":"7d29a994c72d6fbb5530a2d97b7eaaea7dcd03637611aca43c3cadc82535e84a"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.401177 4741 generic.go:334] "Generic (PLEG): container finished" podID="8f495781-1870-461e-b12d-ddf347b0a73c" containerID="829444b757a521955e515b75492ea996bc9c06ad3fa6507cd00d66429d8cb77d" exitCode=0 Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.401208 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5ffsc" event={"ID":"8f495781-1870-461e-b12d-ddf347b0a73c","Type":"ContainerDied","Data":"829444b757a521955e515b75492ea996bc9c06ad3fa6507cd00d66429d8cb77d"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.401259 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5ffsc" event={"ID":"8f495781-1870-461e-b12d-ddf347b0a73c","Type":"ContainerStarted","Data":"2dc64a66850cb58dccca1fc9350a5e25bbd4083a5f3070933b06d02d3cd5c4df"} Sep 29 19:28:40 crc kubenswrapper[4741]: I0929 19:28:40.417300 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-74d7c6f55-9trnx" podStartSLOduration=8.417284272 podStartE2EDuration="8.417284272s" podCreationTimestamp="2025-09-29 19:28:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:40.410264689 +0000 UTC m=+1162.058054021" watchObservedRunningTime="2025-09-29 19:28:40.417284272 +0000 UTC m=+1162.065073604" Sep 29 19:28:41 crc kubenswrapper[4741]: I0929 19:28:41.097145 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3390cf5f-a926-4ea3-9fae-d560cd99789a" path="/var/lib/kubelet/pods/3390cf5f-a926-4ea3-9fae-d560cd99789a/volumes" Sep 29 19:28:41 crc kubenswrapper[4741]: I0929 19:28:41.130467 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:28:41 crc kubenswrapper[4741]: I0929 19:28:41.427772 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerStarted","Data":"80f77bda839b108478ce6706faee0c54b9d3bc47b08c7900f50ee07d24b63772"} Sep 29 19:28:41 crc kubenswrapper[4741]: I0929 19:28:41.898139 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2nvjd" Sep 29 19:28:41 crc kubenswrapper[4741]: I0929 19:28:41.905988 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gcx8j" Sep 29 19:28:41 crc kubenswrapper[4741]: I0929 19:28:41.919491 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5ffsc" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.089949 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pn5c\" (UniqueName: \"kubernetes.io/projected/8f495781-1870-461e-b12d-ddf347b0a73c-kube-api-access-9pn5c\") pod \"8f495781-1870-461e-b12d-ddf347b0a73c\" (UID: \"8f495781-1870-461e-b12d-ddf347b0a73c\") " Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.090652 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjxjp\" (UniqueName: \"kubernetes.io/projected/ce0c5a54-c02f-4f97-a0ea-54383f8b792c-kube-api-access-zjxjp\") pod \"ce0c5a54-c02f-4f97-a0ea-54383f8b792c\" (UID: \"ce0c5a54-c02f-4f97-a0ea-54383f8b792c\") " Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.090715 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghzr2\" (UniqueName: \"kubernetes.io/projected/4f1730aa-2bc9-4005-88fb-4634745de0aa-kube-api-access-ghzr2\") pod \"4f1730aa-2bc9-4005-88fb-4634745de0aa\" (UID: \"4f1730aa-2bc9-4005-88fb-4634745de0aa\") " Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.096375 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f495781-1870-461e-b12d-ddf347b0a73c-kube-api-access-9pn5c" (OuterVolumeSpecName: "kube-api-access-9pn5c") pod "8f495781-1870-461e-b12d-ddf347b0a73c" (UID: "8f495781-1870-461e-b12d-ddf347b0a73c"). InnerVolumeSpecName "kube-api-access-9pn5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.096885 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f1730aa-2bc9-4005-88fb-4634745de0aa-kube-api-access-ghzr2" (OuterVolumeSpecName: "kube-api-access-ghzr2") pod "4f1730aa-2bc9-4005-88fb-4634745de0aa" (UID: "4f1730aa-2bc9-4005-88fb-4634745de0aa"). InnerVolumeSpecName "kube-api-access-ghzr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.099566 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce0c5a54-c02f-4f97-a0ea-54383f8b792c-kube-api-access-zjxjp" (OuterVolumeSpecName: "kube-api-access-zjxjp") pod "ce0c5a54-c02f-4f97-a0ea-54383f8b792c" (UID: "ce0c5a54-c02f-4f97-a0ea-54383f8b792c"). InnerVolumeSpecName "kube-api-access-zjxjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.192440 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pn5c\" (UniqueName: \"kubernetes.io/projected/8f495781-1870-461e-b12d-ddf347b0a73c-kube-api-access-9pn5c\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.192472 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjxjp\" (UniqueName: \"kubernetes.io/projected/ce0c5a54-c02f-4f97-a0ea-54383f8b792c-kube-api-access-zjxjp\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.192486 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghzr2\" (UniqueName: \"kubernetes.io/projected/4f1730aa-2bc9-4005-88fb-4634745de0aa-kube-api-access-ghzr2\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.439029 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5ffsc" event={"ID":"8f495781-1870-461e-b12d-ddf347b0a73c","Type":"ContainerDied","Data":"2dc64a66850cb58dccca1fc9350a5e25bbd4083a5f3070933b06d02d3cd5c4df"} Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.439069 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dc64a66850cb58dccca1fc9350a5e25bbd4083a5f3070933b06d02d3cd5c4df" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.439121 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5ffsc" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.443224 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2nvjd" event={"ID":"4f1730aa-2bc9-4005-88fb-4634745de0aa","Type":"ContainerDied","Data":"ee1810932b65cb9ea5ea50cf322354a512b744e2515658f29f60ac86850c83c9"} Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.443270 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee1810932b65cb9ea5ea50cf322354a512b744e2515658f29f60ac86850c83c9" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.443354 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2nvjd" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.448671 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerStarted","Data":"581c4fd1f8976d3a66c687d4b07bd5749a5f436c29918aaac88161f8a5836335"} Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.451996 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gcx8j" event={"ID":"ce0c5a54-c02f-4f97-a0ea-54383f8b792c","Type":"ContainerDied","Data":"7d29a994c72d6fbb5530a2d97b7eaaea7dcd03637611aca43c3cadc82535e84a"} Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.452084 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d29a994c72d6fbb5530a2d97b7eaaea7dcd03637611aca43c3cadc82535e84a" Sep 29 19:28:42 crc kubenswrapper[4741]: I0929 19:28:42.452117 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gcx8j" Sep 29 19:28:43 crc kubenswrapper[4741]: I0929 19:28:43.462089 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerStarted","Data":"ab0f80c37c03f9f6b81f51816be61bffe6e9f557caa32725c1866d54590aac16"} Sep 29 19:28:44 crc kubenswrapper[4741]: I0929 19:28:44.476335 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerStarted","Data":"3349d6882e6c31fff58ffef007601716946dce1763cc99fd10a3806d00825c94"} Sep 29 19:28:44 crc kubenswrapper[4741]: I0929 19:28:44.477040 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 19:28:44 crc kubenswrapper[4741]: I0929 19:28:44.476549 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-notification-agent" containerID="cri-o://581c4fd1f8976d3a66c687d4b07bd5749a5f436c29918aaac88161f8a5836335" gracePeriod=30 Sep 29 19:28:44 crc kubenswrapper[4741]: I0929 19:28:44.476577 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="sg-core" containerID="cri-o://ab0f80c37c03f9f6b81f51816be61bffe6e9f557caa32725c1866d54590aac16" gracePeriod=30 Sep 29 19:28:44 crc kubenswrapper[4741]: I0929 19:28:44.476608 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="proxy-httpd" containerID="cri-o://3349d6882e6c31fff58ffef007601716946dce1763cc99fd10a3806d00825c94" gracePeriod=30 Sep 29 19:28:44 crc kubenswrapper[4741]: I0929 19:28:44.476480 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-central-agent" containerID="cri-o://80f77bda839b108478ce6706faee0c54b9d3bc47b08c7900f50ee07d24b63772" gracePeriod=30 Sep 29 19:28:44 crc kubenswrapper[4741]: I0929 19:28:44.503100 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.632139246 podStartE2EDuration="5.503078129s" podCreationTimestamp="2025-09-29 19:28:39 +0000 UTC" firstStartedPulling="2025-09-29 19:28:40.382556983 +0000 UTC m=+1162.030346315" lastFinishedPulling="2025-09-29 19:28:44.253495866 +0000 UTC m=+1165.901285198" observedRunningTime="2025-09-29 19:28:44.494923181 +0000 UTC m=+1166.142712513" watchObservedRunningTime="2025-09-29 19:28:44.503078129 +0000 UTC m=+1166.150867461" Sep 29 19:28:45 crc kubenswrapper[4741]: I0929 19:28:45.488718 4741 generic.go:334] "Generic (PLEG): container finished" podID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerID="ab0f80c37c03f9f6b81f51816be61bffe6e9f557caa32725c1866d54590aac16" exitCode=2 Sep 29 19:28:45 crc kubenswrapper[4741]: I0929 19:28:45.488770 4741 generic.go:334] "Generic (PLEG): container finished" podID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerID="581c4fd1f8976d3a66c687d4b07bd5749a5f436c29918aaac88161f8a5836335" exitCode=0 Sep 29 19:28:45 crc kubenswrapper[4741]: I0929 19:28:45.488793 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerDied","Data":"ab0f80c37c03f9f6b81f51816be61bffe6e9f557caa32725c1866d54590aac16"} Sep 29 19:28:45 crc kubenswrapper[4741]: I0929 19:28:45.488820 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerDied","Data":"581c4fd1f8976d3a66c687d4b07bd5749a5f436c29918aaac88161f8a5836335"} Sep 29 19:28:47 crc kubenswrapper[4741]: I0929 19:28:47.506201 4741 generic.go:334] "Generic (PLEG): container finished" podID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerID="80f77bda839b108478ce6706faee0c54b9d3bc47b08c7900f50ee07d24b63772" exitCode=0 Sep 29 19:28:47 crc kubenswrapper[4741]: I0929 19:28:47.506273 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerDied","Data":"80f77bda839b108478ce6706faee0c54b9d3bc47b08c7900f50ee07d24b63772"} Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.298617 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.301310 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.730280 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7d7d-account-create-wwqpm"] Sep 29 19:28:48 crc kubenswrapper[4741]: E0929 19:28:48.730728 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0c5a54-c02f-4f97-a0ea-54383f8b792c" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.730742 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0c5a54-c02f-4f97-a0ea-54383f8b792c" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: E0929 19:28:48.730770 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f495781-1870-461e-b12d-ddf347b0a73c" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.730778 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f495781-1870-461e-b12d-ddf347b0a73c" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: E0929 19:28:48.730800 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1730aa-2bc9-4005-88fb-4634745de0aa" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.730809 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1730aa-2bc9-4005-88fb-4634745de0aa" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.731010 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f495781-1870-461e-b12d-ddf347b0a73c" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.731036 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce0c5a54-c02f-4f97-a0ea-54383f8b792c" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.731060 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1730aa-2bc9-4005-88fb-4634745de0aa" containerName="mariadb-database-create" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.731797 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d7d-account-create-wwqpm" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.737265 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.747120 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d7d-account-create-wwqpm"] Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.917842 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9l9d\" (UniqueName: \"kubernetes.io/projected/08063773-12db-4af5-bf43-6ff10b1672f7-kube-api-access-n9l9d\") pod \"nova-api-7d7d-account-create-wwqpm\" (UID: \"08063773-12db-4af5-bf43-6ff10b1672f7\") " pod="openstack/nova-api-7d7d-account-create-wwqpm" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.923504 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2245-account-create-t9wn8"] Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.924631 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2245-account-create-t9wn8" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.927921 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Sep 29 19:28:48 crc kubenswrapper[4741]: I0929 19:28:48.938559 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2245-account-create-t9wn8"] Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.020116 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9l9d\" (UniqueName: \"kubernetes.io/projected/08063773-12db-4af5-bf43-6ff10b1672f7-kube-api-access-n9l9d\") pod \"nova-api-7d7d-account-create-wwqpm\" (UID: \"08063773-12db-4af5-bf43-6ff10b1672f7\") " pod="openstack/nova-api-7d7d-account-create-wwqpm" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.044949 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9l9d\" (UniqueName: \"kubernetes.io/projected/08063773-12db-4af5-bf43-6ff10b1672f7-kube-api-access-n9l9d\") pod \"nova-api-7d7d-account-create-wwqpm\" (UID: \"08063773-12db-4af5-bf43-6ff10b1672f7\") " pod="openstack/nova-api-7d7d-account-create-wwqpm" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.058208 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d7d-account-create-wwqpm" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.122283 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwd5v\" (UniqueName: \"kubernetes.io/projected/f9e005f0-c926-4dee-a78a-2d5da494bd93-kube-api-access-vwd5v\") pod \"nova-cell0-2245-account-create-t9wn8\" (UID: \"f9e005f0-c926-4dee-a78a-2d5da494bd93\") " pod="openstack/nova-cell0-2245-account-create-t9wn8" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.126054 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5b24-account-create-bhddd"] Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.127354 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b24-account-create-bhddd" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.131777 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.158288 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5b24-account-create-bhddd"] Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.225480 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwd5v\" (UniqueName: \"kubernetes.io/projected/f9e005f0-c926-4dee-a78a-2d5da494bd93-kube-api-access-vwd5v\") pod \"nova-cell0-2245-account-create-t9wn8\" (UID: \"f9e005f0-c926-4dee-a78a-2d5da494bd93\") " pod="openstack/nova-cell0-2245-account-create-t9wn8" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.225558 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks5m8\" (UniqueName: \"kubernetes.io/projected/4d0b8a94-1812-451f-9800-c46d88559954-kube-api-access-ks5m8\") pod \"nova-cell1-5b24-account-create-bhddd\" (UID: \"4d0b8a94-1812-451f-9800-c46d88559954\") " pod="openstack/nova-cell1-5b24-account-create-bhddd" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.249203 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwd5v\" (UniqueName: \"kubernetes.io/projected/f9e005f0-c926-4dee-a78a-2d5da494bd93-kube-api-access-vwd5v\") pod \"nova-cell0-2245-account-create-t9wn8\" (UID: \"f9e005f0-c926-4dee-a78a-2d5da494bd93\") " pod="openstack/nova-cell0-2245-account-create-t9wn8" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.290616 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2245-account-create-t9wn8" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.327130 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks5m8\" (UniqueName: \"kubernetes.io/projected/4d0b8a94-1812-451f-9800-c46d88559954-kube-api-access-ks5m8\") pod \"nova-cell1-5b24-account-create-bhddd\" (UID: \"4d0b8a94-1812-451f-9800-c46d88559954\") " pod="openstack/nova-cell1-5b24-account-create-bhddd" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.353724 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.353949 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-log" containerID="cri-o://3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87" gracePeriod=30 Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.354076 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-httpd" containerID="cri-o://38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0" gracePeriod=30 Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.354131 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks5m8\" (UniqueName: \"kubernetes.io/projected/4d0b8a94-1812-451f-9800-c46d88559954-kube-api-access-ks5m8\") pod \"nova-cell1-5b24-account-create-bhddd\" (UID: \"4d0b8a94-1812-451f-9800-c46d88559954\") " pod="openstack/nova-cell1-5b24-account-create-bhddd" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.515659 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b24-account-create-bhddd" Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.525986 4741 generic.go:334] "Generic (PLEG): container finished" podID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerID="3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87" exitCode=143 Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.526043 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d68f4aea-7345-4129-b25d-8d04773f9ebd","Type":"ContainerDied","Data":"3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87"} Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.601257 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d7d-account-create-wwqpm"] Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.760069 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2245-account-create-t9wn8"] Sep 29 19:28:49 crc kubenswrapper[4741]: W0929 19:28:49.777990 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9e005f0_c926_4dee_a78a_2d5da494bd93.slice/crio-a25879b66d251ef8ff9e7771e8ab71388778f98952bc30172311a0527eb07abd WatchSource:0}: Error finding container a25879b66d251ef8ff9e7771e8ab71388778f98952bc30172311a0527eb07abd: Status 404 returned error can't find the container with id a25879b66d251ef8ff9e7771e8ab71388778f98952bc30172311a0527eb07abd Sep 29 19:28:49 crc kubenswrapper[4741]: I0929 19:28:49.982219 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5b24-account-create-bhddd"] Sep 29 19:28:49 crc kubenswrapper[4741]: W0929 19:28:49.982630 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d0b8a94_1812_451f_9800_c46d88559954.slice/crio-abb474a2e4b8dc318acbff14dca028ef8433a00a5a32ba910b8ad2eac5c9cbf9 WatchSource:0}: Error finding container abb474a2e4b8dc318acbff14dca028ef8433a00a5a32ba910b8ad2eac5c9cbf9: Status 404 returned error can't find the container with id abb474a2e4b8dc318acbff14dca028ef8433a00a5a32ba910b8ad2eac5c9cbf9 Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.535225 4741 generic.go:334] "Generic (PLEG): container finished" podID="f9e005f0-c926-4dee-a78a-2d5da494bd93" containerID="aaae2fd3fe24f56e30fb6d8a9846ff321b4277473b96cefb25838ece71c0bd8f" exitCode=0 Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.535274 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2245-account-create-t9wn8" event={"ID":"f9e005f0-c926-4dee-a78a-2d5da494bd93","Type":"ContainerDied","Data":"aaae2fd3fe24f56e30fb6d8a9846ff321b4277473b96cefb25838ece71c0bd8f"} Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.535596 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2245-account-create-t9wn8" event={"ID":"f9e005f0-c926-4dee-a78a-2d5da494bd93","Type":"ContainerStarted","Data":"a25879b66d251ef8ff9e7771e8ab71388778f98952bc30172311a0527eb07abd"} Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.537276 4741 generic.go:334] "Generic (PLEG): container finished" podID="4d0b8a94-1812-451f-9800-c46d88559954" containerID="3f85075833d16e2d3f3491a5e4ebd1422fbd5759d4cfda248972d88804bd4bf9" exitCode=0 Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.537347 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b24-account-create-bhddd" event={"ID":"4d0b8a94-1812-451f-9800-c46d88559954","Type":"ContainerDied","Data":"3f85075833d16e2d3f3491a5e4ebd1422fbd5759d4cfda248972d88804bd4bf9"} Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.537383 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b24-account-create-bhddd" event={"ID":"4d0b8a94-1812-451f-9800-c46d88559954","Type":"ContainerStarted","Data":"abb474a2e4b8dc318acbff14dca028ef8433a00a5a32ba910b8ad2eac5c9cbf9"} Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.539109 4741 generic.go:334] "Generic (PLEG): container finished" podID="08063773-12db-4af5-bf43-6ff10b1672f7" containerID="9bac786058172af76d1feab8bbd593df29dab72cbf3ab45d0f20758962fff82c" exitCode=0 Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.539152 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d7d-account-create-wwqpm" event={"ID":"08063773-12db-4af5-bf43-6ff10b1672f7","Type":"ContainerDied","Data":"9bac786058172af76d1feab8bbd593df29dab72cbf3ab45d0f20758962fff82c"} Sep 29 19:28:50 crc kubenswrapper[4741]: I0929 19:28:50.539180 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d7d-account-create-wwqpm" event={"ID":"08063773-12db-4af5-bf43-6ff10b1672f7","Type":"ContainerStarted","Data":"7a0c56a78b53b7cb67cb721ec7fa70b509902d4901093159285b70abe12398ae"} Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.234601 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.234831 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-log" containerID="cri-o://cb9ae15c067183744cd6dc00963bb2688701cfa6bf3cfda30d99cddb222a8a6e" gracePeriod=30 Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.235237 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-httpd" containerID="cri-o://49827be4e3437ca826ee30ca5dafe5dba420363ceccdaa1f5c0f8ad732401de7" gracePeriod=30 Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.550204 4741 generic.go:334] "Generic (PLEG): container finished" podID="965471c2-5bf3-441c-a9de-1df141e5641e" containerID="cb9ae15c067183744cd6dc00963bb2688701cfa6bf3cfda30d99cddb222a8a6e" exitCode=143 Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.550356 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965471c2-5bf3-441c-a9de-1df141e5641e","Type":"ContainerDied","Data":"cb9ae15c067183744cd6dc00963bb2688701cfa6bf3cfda30d99cddb222a8a6e"} Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.961201 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2245-account-create-t9wn8" Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.968161 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b24-account-create-bhddd" Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.973873 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d7d-account-create-wwqpm" Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.986902 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwd5v\" (UniqueName: \"kubernetes.io/projected/f9e005f0-c926-4dee-a78a-2d5da494bd93-kube-api-access-vwd5v\") pod \"f9e005f0-c926-4dee-a78a-2d5da494bd93\" (UID: \"f9e005f0-c926-4dee-a78a-2d5da494bd93\") " Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.987099 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks5m8\" (UniqueName: \"kubernetes.io/projected/4d0b8a94-1812-451f-9800-c46d88559954-kube-api-access-ks5m8\") pod \"4d0b8a94-1812-451f-9800-c46d88559954\" (UID: \"4d0b8a94-1812-451f-9800-c46d88559954\") " Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.987120 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9l9d\" (UniqueName: \"kubernetes.io/projected/08063773-12db-4af5-bf43-6ff10b1672f7-kube-api-access-n9l9d\") pod \"08063773-12db-4af5-bf43-6ff10b1672f7\" (UID: \"08063773-12db-4af5-bf43-6ff10b1672f7\") " Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.993769 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d0b8a94-1812-451f-9800-c46d88559954-kube-api-access-ks5m8" (OuterVolumeSpecName: "kube-api-access-ks5m8") pod "4d0b8a94-1812-451f-9800-c46d88559954" (UID: "4d0b8a94-1812-451f-9800-c46d88559954"). InnerVolumeSpecName "kube-api-access-ks5m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:51 crc kubenswrapper[4741]: I0929 19:28:51.995031 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08063773-12db-4af5-bf43-6ff10b1672f7-kube-api-access-n9l9d" (OuterVolumeSpecName: "kube-api-access-n9l9d") pod "08063773-12db-4af5-bf43-6ff10b1672f7" (UID: "08063773-12db-4af5-bf43-6ff10b1672f7"). InnerVolumeSpecName "kube-api-access-n9l9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.003881 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9e005f0-c926-4dee-a78a-2d5da494bd93-kube-api-access-vwd5v" (OuterVolumeSpecName: "kube-api-access-vwd5v") pod "f9e005f0-c926-4dee-a78a-2d5da494bd93" (UID: "f9e005f0-c926-4dee-a78a-2d5da494bd93"). InnerVolumeSpecName "kube-api-access-vwd5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.088869 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks5m8\" (UniqueName: \"kubernetes.io/projected/4d0b8a94-1812-451f-9800-c46d88559954-kube-api-access-ks5m8\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.088907 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9l9d\" (UniqueName: \"kubernetes.io/projected/08063773-12db-4af5-bf43-6ff10b1672f7-kube-api-access-n9l9d\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.088919 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwd5v\" (UniqueName: \"kubernetes.io/projected/f9e005f0-c926-4dee-a78a-2d5da494bd93-kube-api-access-vwd5v\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.560799 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2245-account-create-t9wn8" event={"ID":"f9e005f0-c926-4dee-a78a-2d5da494bd93","Type":"ContainerDied","Data":"a25879b66d251ef8ff9e7771e8ab71388778f98952bc30172311a0527eb07abd"} Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.560860 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a25879b66d251ef8ff9e7771e8ab71388778f98952bc30172311a0527eb07abd" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.560825 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2245-account-create-t9wn8" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.562515 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b24-account-create-bhddd" event={"ID":"4d0b8a94-1812-451f-9800-c46d88559954","Type":"ContainerDied","Data":"abb474a2e4b8dc318acbff14dca028ef8433a00a5a32ba910b8ad2eac5c9cbf9"} Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.562552 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abb474a2e4b8dc318acbff14dca028ef8433a00a5a32ba910b8ad2eac5c9cbf9" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.562605 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b24-account-create-bhddd" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.567243 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d7d-account-create-wwqpm" event={"ID":"08063773-12db-4af5-bf43-6ff10b1672f7","Type":"ContainerDied","Data":"7a0c56a78b53b7cb67cb721ec7fa70b509902d4901093159285b70abe12398ae"} Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.567285 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a0c56a78b53b7cb67cb721ec7fa70b509902d4901093159285b70abe12398ae" Sep 29 19:28:52 crc kubenswrapper[4741]: I0929 19:28:52.567290 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d7d-account-create-wwqpm" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.050248 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.104991 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-logs\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105222 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105333 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj8dv\" (UniqueName: \"kubernetes.io/projected/d68f4aea-7345-4129-b25d-8d04773f9ebd-kube-api-access-qj8dv\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105444 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-httpd-run\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105517 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-scripts\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105635 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-logs" (OuterVolumeSpecName: "logs") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105710 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-combined-ca-bundle\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105835 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-config-data\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105878 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.105989 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-public-tls-certs\") pod \"d68f4aea-7345-4129-b25d-8d04773f9ebd\" (UID: \"d68f4aea-7345-4129-b25d-8d04773f9ebd\") " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.106621 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.106695 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d68f4aea-7345-4129-b25d-8d04773f9ebd-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.110587 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d68f4aea-7345-4129-b25d-8d04773f9ebd-kube-api-access-qj8dv" (OuterVolumeSpecName: "kube-api-access-qj8dv") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "kube-api-access-qj8dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.110935 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.117595 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-scripts" (OuterVolumeSpecName: "scripts") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.144858 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.171687 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-config-data" (OuterVolumeSpecName: "config-data") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.171715 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d68f4aea-7345-4129-b25d-8d04773f9ebd" (UID: "d68f4aea-7345-4129-b25d-8d04773f9ebd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.207682 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.207782 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.207807 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj8dv\" (UniqueName: \"kubernetes.io/projected/d68f4aea-7345-4129-b25d-8d04773f9ebd-kube-api-access-qj8dv\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.207843 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.207856 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.207869 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d68f4aea-7345-4129-b25d-8d04773f9ebd-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.227704 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.309412 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.578008 4741 generic.go:334] "Generic (PLEG): container finished" podID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerID="38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0" exitCode=0 Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.578056 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d68f4aea-7345-4129-b25d-8d04773f9ebd","Type":"ContainerDied","Data":"38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0"} Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.578092 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d68f4aea-7345-4129-b25d-8d04773f9ebd","Type":"ContainerDied","Data":"763617a7a4685e1563da99332b0ae71a7874bd46cabfe82b8a58f983d346fea7"} Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.578117 4741 scope.go:117] "RemoveContainer" containerID="38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.578139 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.606756 4741 scope.go:117] "RemoveContainer" containerID="3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.628179 4741 scope.go:117] "RemoveContainer" containerID="38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.628271 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:28:53 crc kubenswrapper[4741]: E0929 19:28:53.628878 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0\": container with ID starting with 38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0 not found: ID does not exist" containerID="38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.628913 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0"} err="failed to get container status \"38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0\": rpc error: code = NotFound desc = could not find container \"38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0\": container with ID starting with 38946e62a9642a6c9f7961222f5195475f83c1ebe187979a96fc21850b09d1d0 not found: ID does not exist" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.628937 4741 scope.go:117] "RemoveContainer" containerID="3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87" Sep 29 19:28:53 crc kubenswrapper[4741]: E0929 19:28:53.629255 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87\": container with ID starting with 3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87 not found: ID does not exist" containerID="3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.629277 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87"} err="failed to get container status \"3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87\": rpc error: code = NotFound desc = could not find container \"3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87\": container with ID starting with 3269a2354012c27acf5845afa59373d8d15c0313f112034591e09067bc5ebd87 not found: ID does not exist" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.636846 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.654047 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:28:53 crc kubenswrapper[4741]: E0929 19:28:53.654579 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-log" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.654596 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-log" Sep 29 19:28:53 crc kubenswrapper[4741]: E0929 19:28:53.654755 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08063773-12db-4af5-bf43-6ff10b1672f7" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.654767 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="08063773-12db-4af5-bf43-6ff10b1672f7" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: E0929 19:28:53.654780 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-httpd" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.654787 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-httpd" Sep 29 19:28:53 crc kubenswrapper[4741]: E0929 19:28:53.654807 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d0b8a94-1812-451f-9800-c46d88559954" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.654813 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d0b8a94-1812-451f-9800-c46d88559954" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: E0929 19:28:53.654821 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e005f0-c926-4dee-a78a-2d5da494bd93" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.654827 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e005f0-c926-4dee-a78a-2d5da494bd93" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.655084 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-log" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.655108 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" containerName="glance-httpd" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.655154 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9e005f0-c926-4dee-a78a-2d5da494bd93" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.655165 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d0b8a94-1812-451f-9800-c46d88559954" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.655174 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="08063773-12db-4af5-bf43-6ff10b1672f7" containerName="mariadb-account-create" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.656183 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.658126 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.658645 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.672716 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816520 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816600 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816682 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816751 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816791 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-logs\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816835 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816907 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.816964 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj2t8\" (UniqueName: \"kubernetes.io/projected/a0ade550-0eaf-4518-8696-14daf3f034d4-kube-api-access-nj2t8\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.919799 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.919859 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.919903 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.919964 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.919999 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-logs\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.920034 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.920081 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.920120 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj2t8\" (UniqueName: \"kubernetes.io/projected/a0ade550-0eaf-4518-8696-14daf3f034d4-kube-api-access-nj2t8\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.920734 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.921094 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.921204 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-logs\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.931237 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.944921 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.950448 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj2t8\" (UniqueName: \"kubernetes.io/projected/a0ade550-0eaf-4518-8696-14daf3f034d4-kube-api-access-nj2t8\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.964721 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:53 crc kubenswrapper[4741]: I0929 19:28:53.965770 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.026810 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " pod="openstack/glance-default-external-api-0" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.146487 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b64h2"] Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.147697 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.149947 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.150088 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.150439 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bbdlm" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.164322 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b64h2"] Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.228323 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-config-data\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.228582 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-scripts\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.228673 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.229068 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwmzk\" (UniqueName: \"kubernetes.io/projected/6279162a-35c0-4528-970c-f1130fda4fcc-kube-api-access-bwmzk\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.274471 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.331571 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwmzk\" (UniqueName: \"kubernetes.io/projected/6279162a-35c0-4528-970c-f1130fda4fcc-kube-api-access-bwmzk\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.332104 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-config-data\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.332174 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-scripts\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.332202 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.336828 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.337552 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-config-data\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.339143 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-scripts\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.357140 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwmzk\" (UniqueName: \"kubernetes.io/projected/6279162a-35c0-4528-970c-f1130fda4fcc-kube-api-access-bwmzk\") pod \"nova-cell0-conductor-db-sync-b64h2\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.466651 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.598074 4741 generic.go:334] "Generic (PLEG): container finished" podID="965471c2-5bf3-441c-a9de-1df141e5641e" containerID="49827be4e3437ca826ee30ca5dafe5dba420363ceccdaa1f5c0f8ad732401de7" exitCode=0 Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.598181 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965471c2-5bf3-441c-a9de-1df141e5641e","Type":"ContainerDied","Data":"49827be4e3437ca826ee30ca5dafe5dba420363ceccdaa1f5c0f8ad732401de7"} Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.841018 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:28:54 crc kubenswrapper[4741]: I0929 19:28:54.963550 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.041364 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b64h2"] Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050422 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-internal-tls-certs\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050480 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njxt6\" (UniqueName: \"kubernetes.io/projected/965471c2-5bf3-441c-a9de-1df141e5641e-kube-api-access-njxt6\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050508 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050676 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-scripts\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050723 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-combined-ca-bundle\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050778 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-config-data\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050806 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-httpd-run\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.050845 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-logs\") pod \"965471c2-5bf3-441c-a9de-1df141e5641e\" (UID: \"965471c2-5bf3-441c-a9de-1df141e5641e\") " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.052073 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-logs" (OuterVolumeSpecName: "logs") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.056017 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.069693 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/965471c2-5bf3-441c-a9de-1df141e5641e-kube-api-access-njxt6" (OuterVolumeSpecName: "kube-api-access-njxt6") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "kube-api-access-njxt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.084633 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-scripts" (OuterVolumeSpecName: "scripts") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.088612 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.113528 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d68f4aea-7345-4129-b25d-8d04773f9ebd" path="/var/lib/kubelet/pods/d68f4aea-7345-4129-b25d-8d04773f9ebd/volumes" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.126352 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.152839 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.152881 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965471c2-5bf3-441c-a9de-1df141e5641e-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.152894 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njxt6\" (UniqueName: \"kubernetes.io/projected/965471c2-5bf3-441c-a9de-1df141e5641e-kube-api-access-njxt6\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.152919 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.152931 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.152946 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.155775 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.157509 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-config-data" (OuterVolumeSpecName: "config-data") pod "965471c2-5bf3-441c-a9de-1df141e5641e" (UID: "965471c2-5bf3-441c-a9de-1df141e5641e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.178357 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.254488 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.254521 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.254568 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965471c2-5bf3-441c-a9de-1df141e5641e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.620734 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ade550-0eaf-4518-8696-14daf3f034d4","Type":"ContainerStarted","Data":"e1b73040b554cbd478a51c8b5816e58a0a68a53c90c12ff3ca2f8c2045c814bf"} Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.620777 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ade550-0eaf-4518-8696-14daf3f034d4","Type":"ContainerStarted","Data":"bcbc7131e1c30527aba2565c71c4ad302351c56425a915b89798715ebb84417f"} Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.623163 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b64h2" event={"ID":"6279162a-35c0-4528-970c-f1130fda4fcc","Type":"ContainerStarted","Data":"7530f3fcbb2c8c0ff0b0b09233e645fe79849a41d0ffef5d42cd45197a41e666"} Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.625850 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965471c2-5bf3-441c-a9de-1df141e5641e","Type":"ContainerDied","Data":"4d7b1baeb9848bc71d0c5143c2381678f682f4749160024d689041b63b94411e"} Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.625912 4741 scope.go:117] "RemoveContainer" containerID="49827be4e3437ca826ee30ca5dafe5dba420363ceccdaa1f5c0f8ad732401de7" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.625925 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.649829 4741 scope.go:117] "RemoveContainer" containerID="cb9ae15c067183744cd6dc00963bb2688701cfa6bf3cfda30d99cddb222a8a6e" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.674965 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.683514 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.705174 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:28:55 crc kubenswrapper[4741]: E0929 19:28:55.705572 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-log" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.705585 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-log" Sep 29 19:28:55 crc kubenswrapper[4741]: E0929 19:28:55.705606 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-httpd" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.705612 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-httpd" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.705830 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-httpd" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.705847 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" containerName="glance-log" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.711085 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.713679 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.713719 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.728910 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.768973 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.769031 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.769058 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2tfl\" (UniqueName: \"kubernetes.io/projected/c9251106-ba81-446b-be7b-51c4ac7f3f81-kube-api-access-q2tfl\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.769077 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.769096 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.769139 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.769184 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-logs\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.769263 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.871244 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.871290 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2tfl\" (UniqueName: \"kubernetes.io/projected/c9251106-ba81-446b-be7b-51c4ac7f3f81-kube-api-access-q2tfl\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.871309 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.871634 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.872346 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.872443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-logs\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.872467 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.872569 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.873109 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.873488 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-logs\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.873658 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.876926 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.883898 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.884984 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.889035 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.914354 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2tfl\" (UniqueName: \"kubernetes.io/projected/c9251106-ba81-446b-be7b-51c4ac7f3f81-kube-api-access-q2tfl\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:55 crc kubenswrapper[4741]: I0929 19:28:55.919304 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " pod="openstack/glance-default-internal-api-0" Sep 29 19:28:56 crc kubenswrapper[4741]: I0929 19:28:56.046487 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:28:56 crc kubenswrapper[4741]: I0929 19:28:56.649864 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:28:56 crc kubenswrapper[4741]: I0929 19:28:56.676367 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ade550-0eaf-4518-8696-14daf3f034d4","Type":"ContainerStarted","Data":"c9909a6976e00256c9ff763bed1a932c9df0fd365ff4dabc64990d63e5be6cf4"} Sep 29 19:28:56 crc kubenswrapper[4741]: I0929 19:28:56.707505 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.707483011 podStartE2EDuration="3.707483011s" podCreationTimestamp="2025-09-29 19:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:56.702466233 +0000 UTC m=+1178.350255585" watchObservedRunningTime="2025-09-29 19:28:56.707483011 +0000 UTC m=+1178.355272343" Sep 29 19:28:57 crc kubenswrapper[4741]: I0929 19:28:57.101146 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="965471c2-5bf3-441c-a9de-1df141e5641e" path="/var/lib/kubelet/pods/965471c2-5bf3-441c-a9de-1df141e5641e/volumes" Sep 29 19:28:57 crc kubenswrapper[4741]: I0929 19:28:57.698778 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c9251106-ba81-446b-be7b-51c4ac7f3f81","Type":"ContainerStarted","Data":"59e07c173f297fc52b58154fee10e2757e48b2c11cd92731bb4e2fc98f386bc7"} Sep 29 19:28:57 crc kubenswrapper[4741]: I0929 19:28:57.699131 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c9251106-ba81-446b-be7b-51c4ac7f3f81","Type":"ContainerStarted","Data":"ef2b713f3984fe037901c5d1baa89b3295575d3b2815af686cbb1fa59e043a63"} Sep 29 19:28:58 crc kubenswrapper[4741]: I0929 19:28:58.710697 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c9251106-ba81-446b-be7b-51c4ac7f3f81","Type":"ContainerStarted","Data":"b426eb53e2825c670095e43996bf58ecec4760a4bafd71a4d7304962625a63c7"} Sep 29 19:28:58 crc kubenswrapper[4741]: I0929 19:28:58.736657 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.736629174 podStartE2EDuration="3.736629174s" podCreationTimestamp="2025-09-29 19:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:28:58.726767504 +0000 UTC m=+1180.374556836" watchObservedRunningTime="2025-09-29 19:28:58.736629174 +0000 UTC m=+1180.384418506" Sep 29 19:29:01 crc kubenswrapper[4741]: I0929 19:29:01.738819 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:29:01 crc kubenswrapper[4741]: I0929 19:29:01.739279 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:29:02 crc kubenswrapper[4741]: I0929 19:29:02.751914 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b64h2" event={"ID":"6279162a-35c0-4528-970c-f1130fda4fcc","Type":"ContainerStarted","Data":"7e12e5f4e6b247101d296695c9a7418f70e077c57abdf6fe940893887187da15"} Sep 29 19:29:02 crc kubenswrapper[4741]: I0929 19:29:02.767067 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-b64h2" podStartSLOduration=1.6183074720000001 podStartE2EDuration="8.767054325s" podCreationTimestamp="2025-09-29 19:28:54 +0000 UTC" firstStartedPulling="2025-09-29 19:28:55.084757016 +0000 UTC m=+1176.732546338" lastFinishedPulling="2025-09-29 19:29:02.233503859 +0000 UTC m=+1183.881293191" observedRunningTime="2025-09-29 19:29:02.765163236 +0000 UTC m=+1184.412952568" watchObservedRunningTime="2025-09-29 19:29:02.767054325 +0000 UTC m=+1184.414843647" Sep 29 19:29:04 crc kubenswrapper[4741]: I0929 19:29:04.274990 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 19:29:04 crc kubenswrapper[4741]: I0929 19:29:04.275274 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 19:29:04 crc kubenswrapper[4741]: I0929 19:29:04.303548 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 19:29:04 crc kubenswrapper[4741]: I0929 19:29:04.317242 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 19:29:04 crc kubenswrapper[4741]: I0929 19:29:04.774011 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 19:29:04 crc kubenswrapper[4741]: I0929 19:29:04.774079 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.047460 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.049782 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.086548 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.087957 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.606924 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.652006 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.802814 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:06 crc kubenswrapper[4741]: I0929 19:29:06.802860 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:08 crc kubenswrapper[4741]: I0929 19:29:08.717802 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:08 crc kubenswrapper[4741]: I0929 19:29:08.718269 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 19:29:09 crc kubenswrapper[4741]: I0929 19:29:09.917087 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 29 19:29:12 crc kubenswrapper[4741]: I0929 19:29:12.877433 4741 generic.go:334] "Generic (PLEG): container finished" podID="6279162a-35c0-4528-970c-f1130fda4fcc" containerID="7e12e5f4e6b247101d296695c9a7418f70e077c57abdf6fe940893887187da15" exitCode=0 Sep 29 19:29:12 crc kubenswrapper[4741]: I0929 19:29:12.877495 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b64h2" event={"ID":"6279162a-35c0-4528-970c-f1130fda4fcc","Type":"ContainerDied","Data":"7e12e5f4e6b247101d296695c9a7418f70e077c57abdf6fe940893887187da15"} Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.274656 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.279722 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-config-data\") pod \"6279162a-35c0-4528-970c-f1130fda4fcc\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.279816 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-combined-ca-bundle\") pod \"6279162a-35c0-4528-970c-f1130fda4fcc\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.279856 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwmzk\" (UniqueName: \"kubernetes.io/projected/6279162a-35c0-4528-970c-f1130fda4fcc-kube-api-access-bwmzk\") pod \"6279162a-35c0-4528-970c-f1130fda4fcc\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.279887 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-scripts\") pod \"6279162a-35c0-4528-970c-f1130fda4fcc\" (UID: \"6279162a-35c0-4528-970c-f1130fda4fcc\") " Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.286751 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6279162a-35c0-4528-970c-f1130fda4fcc-kube-api-access-bwmzk" (OuterVolumeSpecName: "kube-api-access-bwmzk") pod "6279162a-35c0-4528-970c-f1130fda4fcc" (UID: "6279162a-35c0-4528-970c-f1130fda4fcc"). InnerVolumeSpecName "kube-api-access-bwmzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.291152 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-scripts" (OuterVolumeSpecName: "scripts") pod "6279162a-35c0-4528-970c-f1130fda4fcc" (UID: "6279162a-35c0-4528-970c-f1130fda4fcc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.325855 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6279162a-35c0-4528-970c-f1130fda4fcc" (UID: "6279162a-35c0-4528-970c-f1130fda4fcc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.331158 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-config-data" (OuterVolumeSpecName: "config-data") pod "6279162a-35c0-4528-970c-f1130fda4fcc" (UID: "6279162a-35c0-4528-970c-f1130fda4fcc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.382337 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.382367 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.382376 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6279162a-35c0-4528-970c-f1130fda4fcc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.382413 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwmzk\" (UniqueName: \"kubernetes.io/projected/6279162a-35c0-4528-970c-f1130fda4fcc-kube-api-access-bwmzk\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.907086 4741 generic.go:334] "Generic (PLEG): container finished" podID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerID="3349d6882e6c31fff58ffef007601716946dce1763cc99fd10a3806d00825c94" exitCode=137 Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.907181 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerDied","Data":"3349d6882e6c31fff58ffef007601716946dce1763cc99fd10a3806d00825c94"} Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.909148 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b64h2" event={"ID":"6279162a-35c0-4528-970c-f1130fda4fcc","Type":"ContainerDied","Data":"7530f3fcbb2c8c0ff0b0b09233e645fe79849a41d0ffef5d42cd45197a41e666"} Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.909246 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7530f3fcbb2c8c0ff0b0b09233e645fe79849a41d0ffef5d42cd45197a41e666" Sep 29 19:29:14 crc kubenswrapper[4741]: I0929 19:29:14.909210 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b64h2" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.020899 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 19:29:15 crc kubenswrapper[4741]: E0929 19:29:15.021346 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6279162a-35c0-4528-970c-f1130fda4fcc" containerName="nova-cell0-conductor-db-sync" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.021369 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6279162a-35c0-4528-970c-f1130fda4fcc" containerName="nova-cell0-conductor-db-sync" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.021621 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6279162a-35c0-4528-970c-f1130fda4fcc" containerName="nova-cell0-conductor-db-sync" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.022480 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.026326 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bbdlm" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.026327 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.036046 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.093303 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.093368 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lqb9\" (UniqueName: \"kubernetes.io/projected/cb846a7f-9be5-4a52-837e-a423f7736e79-kube-api-access-8lqb9\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.093437 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.194990 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.195335 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lqb9\" (UniqueName: \"kubernetes.io/projected/cb846a7f-9be5-4a52-837e-a423f7736e79-kube-api-access-8lqb9\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.195370 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.206642 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.213419 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.214782 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lqb9\" (UniqueName: \"kubernetes.io/projected/cb846a7f-9be5-4a52-837e-a423f7736e79-kube-api-access-8lqb9\") pod \"nova-cell0-conductor-0\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.331991 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.355398 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.398214 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s769m\" (UniqueName: \"kubernetes.io/projected/864c8ce0-8c53-4d62-88fd-635c425dd5d6-kube-api-access-s769m\") pod \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.398367 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-log-httpd\") pod \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.398425 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-scripts\") pod \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.399207 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-run-httpd\") pod \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.399256 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-config-data\") pod \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.399296 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-sg-core-conf-yaml\") pod \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.399788 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "864c8ce0-8c53-4d62-88fd-635c425dd5d6" (UID: "864c8ce0-8c53-4d62-88fd-635c425dd5d6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.399824 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "864c8ce0-8c53-4d62-88fd-635c425dd5d6" (UID: "864c8ce0-8c53-4d62-88fd-635c425dd5d6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.400037 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-combined-ca-bundle\") pod \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\" (UID: \"864c8ce0-8c53-4d62-88fd-635c425dd5d6\") " Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.400694 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.400711 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/864c8ce0-8c53-4d62-88fd-635c425dd5d6-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.404573 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-scripts" (OuterVolumeSpecName: "scripts") pod "864c8ce0-8c53-4d62-88fd-635c425dd5d6" (UID: "864c8ce0-8c53-4d62-88fd-635c425dd5d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.404606 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/864c8ce0-8c53-4d62-88fd-635c425dd5d6-kube-api-access-s769m" (OuterVolumeSpecName: "kube-api-access-s769m") pod "864c8ce0-8c53-4d62-88fd-635c425dd5d6" (UID: "864c8ce0-8c53-4d62-88fd-635c425dd5d6"). InnerVolumeSpecName "kube-api-access-s769m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.424825 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "864c8ce0-8c53-4d62-88fd-635c425dd5d6" (UID: "864c8ce0-8c53-4d62-88fd-635c425dd5d6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.482557 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "864c8ce0-8c53-4d62-88fd-635c425dd5d6" (UID: "864c8ce0-8c53-4d62-88fd-635c425dd5d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.498424 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-config-data" (OuterVolumeSpecName: "config-data") pod "864c8ce0-8c53-4d62-88fd-635c425dd5d6" (UID: "864c8ce0-8c53-4d62-88fd-635c425dd5d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.504345 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.504372 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.504383 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.504405 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c8ce0-8c53-4d62-88fd-635c425dd5d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.504415 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s769m\" (UniqueName: \"kubernetes.io/projected/864c8ce0-8c53-4d62-88fd-635c425dd5d6-kube-api-access-s769m\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:15 crc kubenswrapper[4741]: W0929 19:29:15.779119 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb846a7f_9be5_4a52_837e_a423f7736e79.slice/crio-179d9546cedc9ade539f8578cc2b0182f784899b30f3f6e0935dd69d19864d5d WatchSource:0}: Error finding container 179d9546cedc9ade539f8578cc2b0182f784899b30f3f6e0935dd69d19864d5d: Status 404 returned error can't find the container with id 179d9546cedc9ade539f8578cc2b0182f784899b30f3f6e0935dd69d19864d5d Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.780271 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.921187 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cb846a7f-9be5-4a52-837e-a423f7736e79","Type":"ContainerStarted","Data":"179d9546cedc9ade539f8578cc2b0182f784899b30f3f6e0935dd69d19864d5d"} Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.923292 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"864c8ce0-8c53-4d62-88fd-635c425dd5d6","Type":"ContainerDied","Data":"54c8f7a3cd287419b7da5b9edf1fb0a25514af86b30666d927b5ecf11dbc4b4c"} Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.923350 4741 scope.go:117] "RemoveContainer" containerID="3349d6882e6c31fff58ffef007601716946dce1763cc99fd10a3806d00825c94" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.923379 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.951604 4741 scope.go:117] "RemoveContainer" containerID="ab0f80c37c03f9f6b81f51816be61bffe6e9f557caa32725c1866d54590aac16" Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.959882 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:15 crc kubenswrapper[4741]: I0929 19:29:15.971188 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.000511 4741 scope.go:117] "RemoveContainer" containerID="581c4fd1f8976d3a66c687d4b07bd5749a5f436c29918aaac88161f8a5836335" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.006514 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:16 crc kubenswrapper[4741]: E0929 19:29:16.007017 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="proxy-httpd" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007040 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="proxy-httpd" Sep 29 19:29:16 crc kubenswrapper[4741]: E0929 19:29:16.007069 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-central-agent" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007079 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-central-agent" Sep 29 19:29:16 crc kubenswrapper[4741]: E0929 19:29:16.007096 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-notification-agent" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007104 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-notification-agent" Sep 29 19:29:16 crc kubenswrapper[4741]: E0929 19:29:16.007134 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="sg-core" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007143 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="sg-core" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007380 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="proxy-httpd" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007416 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-central-agent" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007428 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="ceilometer-notification-agent" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.007457 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" containerName="sg-core" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.011408 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.020993 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.025727 4741 scope.go:117] "RemoveContainer" containerID="80f77bda839b108478ce6706faee0c54b9d3bc47b08c7900f50ee07d24b63772" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.027750 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.027937 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.113194 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.113227 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.113279 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-scripts\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.113325 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-log-httpd\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.113361 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-config-data\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.113377 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-run-httpd\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.113434 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrxrr\" (UniqueName: \"kubernetes.io/projected/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-kube-api-access-zrxrr\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.215352 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.215576 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.215689 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-scripts\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.215795 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-log-httpd\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.215911 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-config-data\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.216278 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-log-httpd\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.216097 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-run-httpd\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.216713 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrxrr\" (UniqueName: \"kubernetes.io/projected/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-kube-api-access-zrxrr\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.216716 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-run-httpd\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.219848 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-scripts\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.220143 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-config-data\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.221301 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.222139 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.234054 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrxrr\" (UniqueName: \"kubernetes.io/projected/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-kube-api-access-zrxrr\") pod \"ceilometer-0\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.335454 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:16 crc kubenswrapper[4741]: W0929 19:29:16.756075 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod760f134b_0672_4cf2_a4c7_0d6ffea1db4a.slice/crio-6f551232eb1334d309414df32f45f281406c2fb233cc52efd4db1fad53115af6 WatchSource:0}: Error finding container 6f551232eb1334d309414df32f45f281406c2fb233cc52efd4db1fad53115af6: Status 404 returned error can't find the container with id 6f551232eb1334d309414df32f45f281406c2fb233cc52efd4db1fad53115af6 Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.757623 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.935166 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cb846a7f-9be5-4a52-837e-a423f7736e79","Type":"ContainerStarted","Data":"e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144"} Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.935255 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.937429 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerStarted","Data":"6f551232eb1334d309414df32f45f281406c2fb233cc52efd4db1fad53115af6"} Sep 29 19:29:16 crc kubenswrapper[4741]: I0929 19:29:16.952424 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.952404774 podStartE2EDuration="2.952404774s" podCreationTimestamp="2025-09-29 19:29:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:16.948345066 +0000 UTC m=+1198.596134398" watchObservedRunningTime="2025-09-29 19:29:16.952404774 +0000 UTC m=+1198.600194116" Sep 29 19:29:17 crc kubenswrapper[4741]: I0929 19:29:17.102868 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="864c8ce0-8c53-4d62-88fd-635c425dd5d6" path="/var/lib/kubelet/pods/864c8ce0-8c53-4d62-88fd-635c425dd5d6/volumes" Sep 29 19:29:17 crc kubenswrapper[4741]: I0929 19:29:17.953773 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerStarted","Data":"5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d"} Sep 29 19:29:18 crc kubenswrapper[4741]: I0929 19:29:18.972036 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerStarted","Data":"b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef"} Sep 29 19:29:19 crc kubenswrapper[4741]: I0929 19:29:19.982714 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerStarted","Data":"5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9"} Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.390116 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.803019 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fmrvg"] Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.804772 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.806978 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.807186 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.815915 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fmrvg"] Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.904971 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-config-data\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.905176 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.905211 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-scripts\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.905303 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf4tq\" (UniqueName: \"kubernetes.io/projected/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-kube-api-access-sf4tq\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.942217 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.944870 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.950090 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 19:29:20 crc kubenswrapper[4741]: I0929 19:29:20.975361 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.004401 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerStarted","Data":"c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891"} Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.006160 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.006199 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-scripts\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.006258 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf4tq\" (UniqueName: \"kubernetes.io/projected/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-kube-api-access-sf4tq\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.006277 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-config-data\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.008015 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.009507 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.015692 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.017994 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.021945 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.029027 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-scripts\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.045730 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-config-data\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.054580 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.058924 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf4tq\" (UniqueName: \"kubernetes.io/projected/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-kube-api-access-sf4tq\") pod \"nova-cell0-cell-mapping-fmrvg\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.063493 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.534351363 podStartE2EDuration="6.063470161s" podCreationTimestamp="2025-09-29 19:29:15 +0000 UTC" firstStartedPulling="2025-09-29 19:29:16.76275606 +0000 UTC m=+1198.410545392" lastFinishedPulling="2025-09-29 19:29:20.291874858 +0000 UTC m=+1201.939664190" observedRunningTime="2025-09-29 19:29:21.048217711 +0000 UTC m=+1202.696007053" watchObservedRunningTime="2025-09-29 19:29:21.063470161 +0000 UTC m=+1202.711259493" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.126073 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.141289 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.149770 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-config-data\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.149827 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-config-data\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.149885 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.149919 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbm9x\" (UniqueName: \"kubernetes.io/projected/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-kube-api-access-gbm9x\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.149959 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz7lx\" (UniqueName: \"kubernetes.io/projected/3ce9ebef-fbff-42be-8f52-3f949618ae3e-kube-api-access-fz7lx\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.150047 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-logs\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.150079 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.155370 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.155546 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.157946 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253232 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253290 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253331 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-config-data\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253352 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-config-data\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253374 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn44d\" (UniqueName: \"kubernetes.io/projected/c152f034-8b8e-44cf-8b05-48da43cd9378-kube-api-access-mn44d\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253420 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbm9x\" (UniqueName: \"kubernetes.io/projected/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-kube-api-access-gbm9x\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253468 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz7lx\" (UniqueName: \"kubernetes.io/projected/3ce9ebef-fbff-42be-8f52-3f949618ae3e-kube-api-access-fz7lx\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253512 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-logs\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.253531 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.254710 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-logs\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.263850 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.264096 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-config-data\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.265587 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-config-data\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.265737 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.277463 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.278984 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbm9x\" (UniqueName: \"kubernetes.io/projected/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-kube-api-access-gbm9x\") pod \"nova-api-0\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.279985 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.281275 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz7lx\" (UniqueName: \"kubernetes.io/projected/3ce9ebef-fbff-42be-8f52-3f949618ae3e-kube-api-access-fz7lx\") pod \"nova-scheduler-0\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.291664 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.300069 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.341354 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9nm5f"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.342898 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357494 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357544 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-config-data\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357618 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357672 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m474\" (UniqueName: \"kubernetes.io/projected/f0900a38-6648-4c90-9458-c4395b446a3a-kube-api-access-2m474\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357701 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357732 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptcxf\" (UniqueName: \"kubernetes.io/projected/f2d27a82-37d8-4edd-8636-e849cbdae6be-kube-api-access-ptcxf\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357754 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357778 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357818 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn44d\" (UniqueName: \"kubernetes.io/projected/c152f034-8b8e-44cf-8b05-48da43cd9378-kube-api-access-mn44d\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357851 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-config\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357887 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d27a82-37d8-4edd-8636-e849cbdae6be-logs\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357915 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.357960 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.358027 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9nm5f"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.366043 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.366155 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.409627 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn44d\" (UniqueName: \"kubernetes.io/projected/c152f034-8b8e-44cf-8b05-48da43cd9378-kube-api-access-mn44d\") pod \"nova-cell1-novncproxy-0\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.458968 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.459030 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.459080 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-config-data\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460212 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460250 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460308 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460344 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m474\" (UniqueName: \"kubernetes.io/projected/f0900a38-6648-4c90-9458-c4395b446a3a-kube-api-access-2m474\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460368 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptcxf\" (UniqueName: \"kubernetes.io/projected/f2d27a82-37d8-4edd-8636-e849cbdae6be-kube-api-access-ptcxf\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460402 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460441 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-config\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460468 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d27a82-37d8-4edd-8636-e849cbdae6be-logs\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.460488 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.461025 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.461241 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.461908 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-config\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.461928 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d27a82-37d8-4edd-8636-e849cbdae6be-logs\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.468063 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.468350 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-config-data\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.479241 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptcxf\" (UniqueName: \"kubernetes.io/projected/f2d27a82-37d8-4edd-8636-e849cbdae6be-kube-api-access-ptcxf\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.480972 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.493169 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m474\" (UniqueName: \"kubernetes.io/projected/f0900a38-6648-4c90-9458-c4395b446a3a-kube-api-access-2m474\") pod \"dnsmasq-dns-757b4f8459-9nm5f\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.504213 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.526050 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.567045 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.659729 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.775775 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fmrvg"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.898547 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8wwcs"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.903761 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.907791 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.908326 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.920652 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8wwcs"] Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.971614 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqhdb\" (UniqueName: \"kubernetes.io/projected/f8909956-17f7-4fa4-805b-e36c4dd756a5-kube-api-access-rqhdb\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.974767 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.974860 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-scripts\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:21 crc kubenswrapper[4741]: I0929 19:29:21.974916 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-config-data\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.026679 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fmrvg" event={"ID":"629809b6-c9b1-4d39-a89c-4d8418e1bdc0","Type":"ContainerStarted","Data":"b7c7a3c2f54b2a585d13e740bce23f3ba4728121ed27cb6741166346468bfaac"} Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.077745 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqhdb\" (UniqueName: \"kubernetes.io/projected/f8909956-17f7-4fa4-805b-e36c4dd756a5-kube-api-access-rqhdb\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.077904 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.077944 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-scripts\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.077981 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-config-data\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.088877 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-scripts\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.089163 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-config-data\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.092310 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.115204 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqhdb\" (UniqueName: \"kubernetes.io/projected/f8909956-17f7-4fa4-805b-e36c4dd756a5-kube-api-access-rqhdb\") pod \"nova-cell1-conductor-db-sync-8wwcs\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.211556 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.249771 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.397551 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:22 crc kubenswrapper[4741]: W0929 19:29:22.409506 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b082f7b_d8b9_4777_80c7_52f8f3b243cb.slice/crio-739413966ca129e79a20a040545df098e42f8153f0d4216cd2fbdc624b6dd9ef WatchSource:0}: Error finding container 739413966ca129e79a20a040545df098e42f8153f0d4216cd2fbdc624b6dd9ef: Status 404 returned error can't find the container with id 739413966ca129e79a20a040545df098e42f8153f0d4216cd2fbdc624b6dd9ef Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.526047 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:22 crc kubenswrapper[4741]: W0929 19:29:22.537687 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2d27a82_37d8_4edd_8636_e849cbdae6be.slice/crio-e4af457941f62729a8f2469874fea76ba5f6ddfac097b9997573719fc48bfa8f WatchSource:0}: Error finding container e4af457941f62729a8f2469874fea76ba5f6ddfac097b9997573719fc48bfa8f: Status 404 returned error can't find the container with id e4af457941f62729a8f2469874fea76ba5f6ddfac097b9997573719fc48bfa8f Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.565838 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.603979 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9nm5f"] Sep 29 19:29:22 crc kubenswrapper[4741]: W0929 19:29:22.624964 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0900a38_6648_4c90_9458_c4395b446a3a.slice/crio-af8964bb39257bb53d73fa5b041b8f800b07703ce84bcfbefc2583cce8891e0d WatchSource:0}: Error finding container af8964bb39257bb53d73fa5b041b8f800b07703ce84bcfbefc2583cce8891e0d: Status 404 returned error can't find the container with id af8964bb39257bb53d73fa5b041b8f800b07703ce84bcfbefc2583cce8891e0d Sep 29 19:29:22 crc kubenswrapper[4741]: I0929 19:29:22.752928 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8wwcs"] Sep 29 19:29:22 crc kubenswrapper[4741]: W0929 19:29:22.755778 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8909956_17f7_4fa4_805b_e36c4dd756a5.slice/crio-1d636e99d2f6f3843c98cb677e8ac7f7cfb46f629835eb9a5c01a9ec7ee9db14 WatchSource:0}: Error finding container 1d636e99d2f6f3843c98cb677e8ac7f7cfb46f629835eb9a5c01a9ec7ee9db14: Status 404 returned error can't find the container with id 1d636e99d2f6f3843c98cb677e8ac7f7cfb46f629835eb9a5c01a9ec7ee9db14 Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.041983 4741 generic.go:334] "Generic (PLEG): container finished" podID="f0900a38-6648-4c90-9458-c4395b446a3a" containerID="d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf" exitCode=0 Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.042025 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" event={"ID":"f0900a38-6648-4c90-9458-c4395b446a3a","Type":"ContainerDied","Data":"d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.042088 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" event={"ID":"f0900a38-6648-4c90-9458-c4395b446a3a","Type":"ContainerStarted","Data":"af8964bb39257bb53d73fa5b041b8f800b07703ce84bcfbefc2583cce8891e0d"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.048279 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c152f034-8b8e-44cf-8b05-48da43cd9378","Type":"ContainerStarted","Data":"fc1ed0ee09a7879f8c1c6a7b9f42420e032e2ba552f050c758e63eab83a92fd3"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.050516 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d27a82-37d8-4edd-8636-e849cbdae6be","Type":"ContainerStarted","Data":"e4af457941f62729a8f2469874fea76ba5f6ddfac097b9997573719fc48bfa8f"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.052128 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ce9ebef-fbff-42be-8f52-3f949618ae3e","Type":"ContainerStarted","Data":"34cd3241853cbc39a8d19d2dafd5236fb51819d9960de42be53dcbca967879bb"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.066762 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fmrvg" event={"ID":"629809b6-c9b1-4d39-a89c-4d8418e1bdc0","Type":"ContainerStarted","Data":"58f9c958476dc4429b27a3cf8d2a755c71d7cd37a2d48d47dc954ccd9f7b2781"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.083775 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b082f7b-d8b9-4777-80c7-52f8f3b243cb","Type":"ContainerStarted","Data":"739413966ca129e79a20a040545df098e42f8153f0d4216cd2fbdc624b6dd9ef"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.098286 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fmrvg" podStartSLOduration=3.098265052 podStartE2EDuration="3.098265052s" podCreationTimestamp="2025-09-29 19:29:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:23.080779032 +0000 UTC m=+1204.728568364" watchObservedRunningTime="2025-09-29 19:29:23.098265052 +0000 UTC m=+1204.746054384" Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.104752 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" event={"ID":"f8909956-17f7-4fa4-805b-e36c4dd756a5","Type":"ContainerStarted","Data":"56f048944583f6a7fea3651342397a37fb4283a7918ed222db829e37ecd52e27"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.104799 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" event={"ID":"f8909956-17f7-4fa4-805b-e36c4dd756a5","Type":"ContainerStarted","Data":"1d636e99d2f6f3843c98cb677e8ac7f7cfb46f629835eb9a5c01a9ec7ee9db14"} Sep 29 19:29:23 crc kubenswrapper[4741]: I0929 19:29:23.117729 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" podStartSLOduration=2.117710433 podStartE2EDuration="2.117710433s" podCreationTimestamp="2025-09-29 19:29:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:23.10555239 +0000 UTC m=+1204.753341722" watchObservedRunningTime="2025-09-29 19:29:23.117710433 +0000 UTC m=+1204.765499765" Sep 29 19:29:24 crc kubenswrapper[4741]: I0929 19:29:24.104938 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" event={"ID":"f0900a38-6648-4c90-9458-c4395b446a3a","Type":"ContainerStarted","Data":"816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11"} Sep 29 19:29:24 crc kubenswrapper[4741]: I0929 19:29:24.129181 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" podStartSLOduration=3.129164407 podStartE2EDuration="3.129164407s" podCreationTimestamp="2025-09-29 19:29:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:24.123506619 +0000 UTC m=+1205.771295961" watchObservedRunningTime="2025-09-29 19:29:24.129164407 +0000 UTC m=+1205.776953739" Sep 29 19:29:24 crc kubenswrapper[4741]: I0929 19:29:24.466461 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:24 crc kubenswrapper[4741]: I0929 19:29:24.514685 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:25 crc kubenswrapper[4741]: I0929 19:29:25.114537 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.124349 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d27a82-37d8-4edd-8636-e849cbdae6be","Type":"ContainerStarted","Data":"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb"} Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.124413 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d27a82-37d8-4edd-8636-e849cbdae6be","Type":"ContainerStarted","Data":"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3"} Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.124470 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-metadata" containerID="cri-o://74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb" gracePeriod=30 Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.124517 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-log" containerID="cri-o://5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3" gracePeriod=30 Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.127268 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ce9ebef-fbff-42be-8f52-3f949618ae3e","Type":"ContainerStarted","Data":"1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634"} Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.131121 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b082f7b-d8b9-4777-80c7-52f8f3b243cb","Type":"ContainerStarted","Data":"8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a"} Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.131164 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b082f7b-d8b9-4777-80c7-52f8f3b243cb","Type":"ContainerStarted","Data":"448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166"} Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.134073 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="c152f034-8b8e-44cf-8b05-48da43cd9378" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2" gracePeriod=30 Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.134326 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c152f034-8b8e-44cf-8b05-48da43cd9378","Type":"ContainerStarted","Data":"4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2"} Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.150637 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.329563605 podStartE2EDuration="5.150619089s" podCreationTimestamp="2025-09-29 19:29:21 +0000 UTC" firstStartedPulling="2025-09-29 19:29:22.54165171 +0000 UTC m=+1204.189441042" lastFinishedPulling="2025-09-29 19:29:25.362707194 +0000 UTC m=+1207.010496526" observedRunningTime="2025-09-29 19:29:26.144819217 +0000 UTC m=+1207.792608549" watchObservedRunningTime="2025-09-29 19:29:26.150619089 +0000 UTC m=+1207.798408421" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.173872 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.213760393 podStartE2EDuration="6.173853789s" podCreationTimestamp="2025-09-29 19:29:20 +0000 UTC" firstStartedPulling="2025-09-29 19:29:22.411459796 +0000 UTC m=+1204.059249128" lastFinishedPulling="2025-09-29 19:29:25.371553192 +0000 UTC m=+1207.019342524" observedRunningTime="2025-09-29 19:29:26.169400959 +0000 UTC m=+1207.817190291" watchObservedRunningTime="2025-09-29 19:29:26.173853789 +0000 UTC m=+1207.821643121" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.219884 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.085231292 podStartE2EDuration="5.219866126s" podCreationTimestamp="2025-09-29 19:29:21 +0000 UTC" firstStartedPulling="2025-09-29 19:29:22.228795523 +0000 UTC m=+1203.876584855" lastFinishedPulling="2025-09-29 19:29:25.363430327 +0000 UTC m=+1207.011219689" observedRunningTime="2025-09-29 19:29:26.195833921 +0000 UTC m=+1207.843623263" watchObservedRunningTime="2025-09-29 19:29:26.219866126 +0000 UTC m=+1207.867655458" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.468669 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.505241 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.661082 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.661124 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:29:26 crc kubenswrapper[4741]: I0929 19:29:26.993943 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.012551 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.221853671 podStartE2EDuration="7.01253003s" podCreationTimestamp="2025-09-29 19:29:20 +0000 UTC" firstStartedPulling="2025-09-29 19:29:22.573401378 +0000 UTC m=+1204.221190700" lastFinishedPulling="2025-09-29 19:29:25.364077727 +0000 UTC m=+1207.011867059" observedRunningTime="2025-09-29 19:29:26.223687087 +0000 UTC m=+1207.871476419" watchObservedRunningTime="2025-09-29 19:29:27.01253003 +0000 UTC m=+1208.660319372" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.110154 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptcxf\" (UniqueName: \"kubernetes.io/projected/f2d27a82-37d8-4edd-8636-e849cbdae6be-kube-api-access-ptcxf\") pod \"f2d27a82-37d8-4edd-8636-e849cbdae6be\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.110415 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-config-data\") pod \"f2d27a82-37d8-4edd-8636-e849cbdae6be\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.110553 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d27a82-37d8-4edd-8636-e849cbdae6be-logs\") pod \"f2d27a82-37d8-4edd-8636-e849cbdae6be\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.110605 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-combined-ca-bundle\") pod \"f2d27a82-37d8-4edd-8636-e849cbdae6be\" (UID: \"f2d27a82-37d8-4edd-8636-e849cbdae6be\") " Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.111423 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2d27a82-37d8-4edd-8636-e849cbdae6be-logs" (OuterVolumeSpecName: "logs") pod "f2d27a82-37d8-4edd-8636-e849cbdae6be" (UID: "f2d27a82-37d8-4edd-8636-e849cbdae6be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.129935 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d27a82-37d8-4edd-8636-e849cbdae6be-kube-api-access-ptcxf" (OuterVolumeSpecName: "kube-api-access-ptcxf") pod "f2d27a82-37d8-4edd-8636-e849cbdae6be" (UID: "f2d27a82-37d8-4edd-8636-e849cbdae6be"). InnerVolumeSpecName "kube-api-access-ptcxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.143933 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-config-data" (OuterVolumeSpecName: "config-data") pod "f2d27a82-37d8-4edd-8636-e849cbdae6be" (UID: "f2d27a82-37d8-4edd-8636-e849cbdae6be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.145794 4741 generic.go:334] "Generic (PLEG): container finished" podID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerID="74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb" exitCode=0 Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.145844 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2d27a82-37d8-4edd-8636-e849cbdae6be" (UID: "f2d27a82-37d8-4edd-8636-e849cbdae6be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.145872 4741 generic.go:334] "Generic (PLEG): container finished" podID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerID="5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3" exitCode=143 Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.145937 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.180969 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d27a82-37d8-4edd-8636-e849cbdae6be","Type":"ContainerDied","Data":"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb"} Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.181013 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d27a82-37d8-4edd-8636-e849cbdae6be","Type":"ContainerDied","Data":"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3"} Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.181040 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d27a82-37d8-4edd-8636-e849cbdae6be","Type":"ContainerDied","Data":"e4af457941f62729a8f2469874fea76ba5f6ddfac097b9997573719fc48bfa8f"} Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.181058 4741 scope.go:117] "RemoveContainer" containerID="74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.213080 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d27a82-37d8-4edd-8636-e849cbdae6be-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.213116 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.213128 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptcxf\" (UniqueName: \"kubernetes.io/projected/f2d27a82-37d8-4edd-8636-e849cbdae6be-kube-api-access-ptcxf\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.213138 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d27a82-37d8-4edd-8636-e849cbdae6be-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.241634 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.249374 4741 scope.go:117] "RemoveContainer" containerID="5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.281139 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.283633 4741 scope.go:117] "RemoveContainer" containerID="74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb" Sep 29 19:29:27 crc kubenswrapper[4741]: E0929 19:29:27.284297 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb\": container with ID starting with 74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb not found: ID does not exist" containerID="74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.284342 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb"} err="failed to get container status \"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb\": rpc error: code = NotFound desc = could not find container \"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb\": container with ID starting with 74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb not found: ID does not exist" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.284372 4741 scope.go:117] "RemoveContainer" containerID="5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3" Sep 29 19:29:27 crc kubenswrapper[4741]: E0929 19:29:27.284840 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3\": container with ID starting with 5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3 not found: ID does not exist" containerID="5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.284869 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3"} err="failed to get container status \"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3\": rpc error: code = NotFound desc = could not find container \"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3\": container with ID starting with 5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3 not found: ID does not exist" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.284891 4741 scope.go:117] "RemoveContainer" containerID="74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.285468 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb"} err="failed to get container status \"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb\": rpc error: code = NotFound desc = could not find container \"74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb\": container with ID starting with 74b71bc481690f5882ae11616ee97e0630f0fa437c52bd4ee17959a082e8a9eb not found: ID does not exist" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.285516 4741 scope.go:117] "RemoveContainer" containerID="5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.286919 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3"} err="failed to get container status \"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3\": rpc error: code = NotFound desc = could not find container \"5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3\": container with ID starting with 5b15e42e4eceee24abeb0045c019399b5a121c15fe2c0f1489a6d24f9f3cf0a3 not found: ID does not exist" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.302858 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:27 crc kubenswrapper[4741]: E0929 19:29:27.303371 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-log" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.303412 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-log" Sep 29 19:29:27 crc kubenswrapper[4741]: E0929 19:29:27.303430 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-metadata" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.303438 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-metadata" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.303744 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-metadata" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.303778 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" containerName="nova-metadata-log" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.305128 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.307871 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.310352 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.323745 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.416652 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.416695 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.416748 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n87q2\" (UniqueName: \"kubernetes.io/projected/182049e3-b96f-4c7e-b737-55375eb23758-kube-api-access-n87q2\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.416795 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182049e3-b96f-4c7e-b737-55375eb23758-logs\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.416825 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-config-data\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.518570 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182049e3-b96f-4c7e-b737-55375eb23758-logs\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.518622 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-config-data\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.518714 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.518737 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.518781 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n87q2\" (UniqueName: \"kubernetes.io/projected/182049e3-b96f-4c7e-b737-55375eb23758-kube-api-access-n87q2\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.519044 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182049e3-b96f-4c7e-b737-55375eb23758-logs\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.535557 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.545644 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.548523 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n87q2\" (UniqueName: \"kubernetes.io/projected/182049e3-b96f-4c7e-b737-55375eb23758-kube-api-access-n87q2\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.553468 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-config-data\") pod \"nova-metadata-0\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " pod="openstack/nova-metadata-0" Sep 29 19:29:27 crc kubenswrapper[4741]: I0929 19:29:27.633010 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:28 crc kubenswrapper[4741]: I0929 19:29:28.079830 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:28 crc kubenswrapper[4741]: I0929 19:29:28.158494 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"182049e3-b96f-4c7e-b737-55375eb23758","Type":"ContainerStarted","Data":"dccb25610f36702dd33d7fdf6125bc01a80230de0e8dccb2db7485ad3aa170f5"} Sep 29 19:29:29 crc kubenswrapper[4741]: I0929 19:29:29.102929 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2d27a82-37d8-4edd-8636-e849cbdae6be" path="/var/lib/kubelet/pods/f2d27a82-37d8-4edd-8636-e849cbdae6be/volumes" Sep 29 19:29:29 crc kubenswrapper[4741]: I0929 19:29:29.172010 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"182049e3-b96f-4c7e-b737-55375eb23758","Type":"ContainerStarted","Data":"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067"} Sep 29 19:29:29 crc kubenswrapper[4741]: I0929 19:29:29.172245 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"182049e3-b96f-4c7e-b737-55375eb23758","Type":"ContainerStarted","Data":"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad"} Sep 29 19:29:30 crc kubenswrapper[4741]: I0929 19:29:30.186368 4741 generic.go:334] "Generic (PLEG): container finished" podID="f8909956-17f7-4fa4-805b-e36c4dd756a5" containerID="56f048944583f6a7fea3651342397a37fb4283a7918ed222db829e37ecd52e27" exitCode=0 Sep 29 19:29:30 crc kubenswrapper[4741]: I0929 19:29:30.186514 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" event={"ID":"f8909956-17f7-4fa4-805b-e36c4dd756a5","Type":"ContainerDied","Data":"56f048944583f6a7fea3651342397a37fb4283a7918ed222db829e37ecd52e27"} Sep 29 19:29:30 crc kubenswrapper[4741]: I0929 19:29:30.191569 4741 generic.go:334] "Generic (PLEG): container finished" podID="629809b6-c9b1-4d39-a89c-4d8418e1bdc0" containerID="58f9c958476dc4429b27a3cf8d2a755c71d7cd37a2d48d47dc954ccd9f7b2781" exitCode=0 Sep 29 19:29:30 crc kubenswrapper[4741]: I0929 19:29:30.192585 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fmrvg" event={"ID":"629809b6-c9b1-4d39-a89c-4d8418e1bdc0","Type":"ContainerDied","Data":"58f9c958476dc4429b27a3cf8d2a755c71d7cd37a2d48d47dc954ccd9f7b2781"} Sep 29 19:29:30 crc kubenswrapper[4741]: I0929 19:29:30.212309 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.212289152 podStartE2EDuration="3.212289152s" podCreationTimestamp="2025-09-29 19:29:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:29.204596997 +0000 UTC m=+1210.852386349" watchObservedRunningTime="2025-09-29 19:29:30.212289152 +0000 UTC m=+1211.860078484" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.468959 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.501611 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.528574 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.567844 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.568078 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.601883 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-m7rm4"] Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.602193 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" podUID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerName="dnsmasq-dns" containerID="cri-o://6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde" gracePeriod=10 Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.713429 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.739815 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.739861 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.739903 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.740851 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1bdfb84a67cbc508e610930001920c7a19e715be8c9f4e424f1f5b91fd396f8"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.740897 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://f1bdfb84a67cbc508e610930001920c7a19e715be8c9f4e424f1f5b91fd396f8" gracePeriod=600 Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.766119 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.802451 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf4tq\" (UniqueName: \"kubernetes.io/projected/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-kube-api-access-sf4tq\") pod \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.802892 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-config-data\") pod \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.802972 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-combined-ca-bundle\") pod \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.803110 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-scripts\") pod \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\" (UID: \"629809b6-c9b1-4d39-a89c-4d8418e1bdc0\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.813806 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-scripts" (OuterVolumeSpecName: "scripts") pod "629809b6-c9b1-4d39-a89c-4d8418e1bdc0" (UID: "629809b6-c9b1-4d39-a89c-4d8418e1bdc0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.813880 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-kube-api-access-sf4tq" (OuterVolumeSpecName: "kube-api-access-sf4tq") pod "629809b6-c9b1-4d39-a89c-4d8418e1bdc0" (UID: "629809b6-c9b1-4d39-a89c-4d8418e1bdc0"). InnerVolumeSpecName "kube-api-access-sf4tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.835965 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "629809b6-c9b1-4d39-a89c-4d8418e1bdc0" (UID: "629809b6-c9b1-4d39-a89c-4d8418e1bdc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.847597 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-config-data" (OuterVolumeSpecName: "config-data") pod "629809b6-c9b1-4d39-a89c-4d8418e1bdc0" (UID: "629809b6-c9b1-4d39-a89c-4d8418e1bdc0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.904924 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-combined-ca-bundle\") pod \"f8909956-17f7-4fa4-805b-e36c4dd756a5\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.905058 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-scripts\") pod \"f8909956-17f7-4fa4-805b-e36c4dd756a5\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.905096 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-config-data\") pod \"f8909956-17f7-4fa4-805b-e36c4dd756a5\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.905129 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqhdb\" (UniqueName: \"kubernetes.io/projected/f8909956-17f7-4fa4-805b-e36c4dd756a5-kube-api-access-rqhdb\") pod \"f8909956-17f7-4fa4-805b-e36c4dd756a5\" (UID: \"f8909956-17f7-4fa4-805b-e36c4dd756a5\") " Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.905749 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.905773 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.905787 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.905798 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf4tq\" (UniqueName: \"kubernetes.io/projected/629809b6-c9b1-4d39-a89c-4d8418e1bdc0-kube-api-access-sf4tq\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.912037 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8909956-17f7-4fa4-805b-e36c4dd756a5-kube-api-access-rqhdb" (OuterVolumeSpecName: "kube-api-access-rqhdb") pod "f8909956-17f7-4fa4-805b-e36c4dd756a5" (UID: "f8909956-17f7-4fa4-805b-e36c4dd756a5"). InnerVolumeSpecName "kube-api-access-rqhdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.912158 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-scripts" (OuterVolumeSpecName: "scripts") pod "f8909956-17f7-4fa4-805b-e36c4dd756a5" (UID: "f8909956-17f7-4fa4-805b-e36c4dd756a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.948500 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8909956-17f7-4fa4-805b-e36c4dd756a5" (UID: "f8909956-17f7-4fa4-805b-e36c4dd756a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:31 crc kubenswrapper[4741]: I0929 19:29:31.948372 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-config-data" (OuterVolumeSpecName: "config-data") pod "f8909956-17f7-4fa4-805b-e36c4dd756a5" (UID: "f8909956-17f7-4fa4-805b-e36c4dd756a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.008624 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.008706 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.008746 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8909956-17f7-4fa4-805b-e36c4dd756a5-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.008757 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqhdb\" (UniqueName: \"kubernetes.io/projected/f8909956-17f7-4fa4-805b-e36c4dd756a5-kube-api-access-rqhdb\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.076175 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.212307 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-sb\") pod \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.212729 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-swift-storage-0\") pod \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.212833 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-svc\") pod \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.212903 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxnj2\" (UniqueName: \"kubernetes.io/projected/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-kube-api-access-jxnj2\") pod \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.213008 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-nb\") pod \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.213045 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-config\") pod \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\" (UID: \"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0\") " Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.217845 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-kube-api-access-jxnj2" (OuterVolumeSpecName: "kube-api-access-jxnj2") pod "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" (UID: "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0"). InnerVolumeSpecName "kube-api-access-jxnj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.220033 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.220100 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8wwcs" event={"ID":"f8909956-17f7-4fa4-805b-e36c4dd756a5","Type":"ContainerDied","Data":"1d636e99d2f6f3843c98cb677e8ac7f7cfb46f629835eb9a5c01a9ec7ee9db14"} Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.220168 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d636e99d2f6f3843c98cb677e8ac7f7cfb46f629835eb9a5c01a9ec7ee9db14" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.223550 4741 generic.go:334] "Generic (PLEG): container finished" podID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerID="6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde" exitCode=0 Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.223650 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" event={"ID":"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0","Type":"ContainerDied","Data":"6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde"} Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.223683 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" event={"ID":"1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0","Type":"ContainerDied","Data":"3256c8ad423d65ea473b4f3da8441b6bf99d3ef8bb5522ebcd79ca61b0561003"} Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.223709 4741 scope.go:117] "RemoveContainer" containerID="6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.223905 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-m7rm4" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.246919 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="f1bdfb84a67cbc508e610930001920c7a19e715be8c9f4e424f1f5b91fd396f8" exitCode=0 Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.247005 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"f1bdfb84a67cbc508e610930001920c7a19e715be8c9f4e424f1f5b91fd396f8"} Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.247076 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"a962e46acb3fe13f6d108a85b06c3220bc7a353adc196e7c6597c33dec71e863"} Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.253403 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fmrvg" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.254277 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fmrvg" event={"ID":"629809b6-c9b1-4d39-a89c-4d8418e1bdc0","Type":"ContainerDied","Data":"b7c7a3c2f54b2a585d13e740bce23f3ba4728121ed27cb6741166346468bfaac"} Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.254642 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7c7a3c2f54b2a585d13e740bce23f3ba4728121ed27cb6741166346468bfaac" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.285188 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" (UID: "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.288908 4741 scope.go:117] "RemoveContainer" containerID="8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.324850 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxnj2\" (UniqueName: \"kubernetes.io/projected/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-kube-api-access-jxnj2\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.324898 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.337971 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.365277 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" (UID: "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.365492 4741 scope.go:117] "RemoveContainer" containerID="6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.365515 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" (UID: "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.365900 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-config" (OuterVolumeSpecName: "config") pod "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" (UID: "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.366543 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" (UID: "1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:29:32 crc kubenswrapper[4741]: E0929 19:29:32.366686 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde\": container with ID starting with 6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde not found: ID does not exist" containerID="6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.366744 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde"} err="failed to get container status \"6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde\": rpc error: code = NotFound desc = could not find container \"6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde\": container with ID starting with 6f6bd10685ba99ec65b379fd05406b6c4c376195924ac09be7848b554fb81bde not found: ID does not exist" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.366769 4741 scope.go:117] "RemoveContainer" containerID="8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb" Sep 29 19:29:32 crc kubenswrapper[4741]: E0929 19:29:32.378989 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb\": container with ID starting with 8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb not found: ID does not exist" containerID="8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.379060 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb"} err="failed to get container status \"8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb\": rpc error: code = NotFound desc = could not find container \"8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb\": container with ID starting with 8677590a89b9f9d1e44180ba9a0d240068ae3293dbafa260d987a0d8cc56cdfb not found: ID does not exist" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.379093 4741 scope.go:117] "RemoveContainer" containerID="d64910008ab65af756a5c6e2b068a8f502ccd9ab9a8a5bb1eac1cdbedf684c46" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.397702 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 19:29:32 crc kubenswrapper[4741]: E0929 19:29:32.398239 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerName="dnsmasq-dns" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.398253 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerName="dnsmasq-dns" Sep 29 19:29:32 crc kubenswrapper[4741]: E0929 19:29:32.398285 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629809b6-c9b1-4d39-a89c-4d8418e1bdc0" containerName="nova-manage" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.398318 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="629809b6-c9b1-4d39-a89c-4d8418e1bdc0" containerName="nova-manage" Sep 29 19:29:32 crc kubenswrapper[4741]: E0929 19:29:32.398334 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerName="init" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.398342 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerName="init" Sep 29 19:29:32 crc kubenswrapper[4741]: E0929 19:29:32.398354 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8909956-17f7-4fa4-805b-e36c4dd756a5" containerName="nova-cell1-conductor-db-sync" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.398360 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8909956-17f7-4fa4-805b-e36c4dd756a5" containerName="nova-cell1-conductor-db-sync" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.398661 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8909956-17f7-4fa4-805b-e36c4dd756a5" containerName="nova-cell1-conductor-db-sync" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.398675 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="629809b6-c9b1-4d39-a89c-4d8418e1bdc0" containerName="nova-manage" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.398717 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" containerName="dnsmasq-dns" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.399434 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.406788 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.409247 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.427940 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.427968 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.427978 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.427986 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.467610 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.467888 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-log" containerID="cri-o://448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166" gracePeriod=30 Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.468310 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-api" containerID="cri-o://8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a" gracePeriod=30 Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.472915 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.179:8774/\": EOF" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.474166 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.179:8774/\": EOF" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.485527 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.485797 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-log" containerID="cri-o://ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad" gracePeriod=30 Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.486061 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-metadata" containerID="cri-o://cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067" gracePeriod=30 Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.529658 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.530013 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f8p4\" (UniqueName: \"kubernetes.io/projected/93c5456d-7a59-4991-ad3d-58b04db78b24-kube-api-access-6f8p4\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.530104 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.631240 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.631479 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f8p4\" (UniqueName: \"kubernetes.io/projected/93c5456d-7a59-4991-ad3d-58b04db78b24-kube-api-access-6f8p4\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.631534 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.635568 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.635601 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.636856 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.637102 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.647999 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f8p4\" (UniqueName: \"kubernetes.io/projected/93c5456d-7a59-4991-ad3d-58b04db78b24-kube-api-access-6f8p4\") pod \"nova-cell1-conductor-0\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.732630 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.837667 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-m7rm4"] Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.844347 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-m7rm4"] Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.903257 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:32 crc kubenswrapper[4741]: I0929 19:29:32.994324 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.100641 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0" path="/var/lib/kubelet/pods/1d404bf3-e8d2-4010-8c59-bc6a6e0c1fb0/volumes" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.142000 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-config-data\") pod \"182049e3-b96f-4c7e-b737-55375eb23758\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.142059 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n87q2\" (UniqueName: \"kubernetes.io/projected/182049e3-b96f-4c7e-b737-55375eb23758-kube-api-access-n87q2\") pod \"182049e3-b96f-4c7e-b737-55375eb23758\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.142082 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182049e3-b96f-4c7e-b737-55375eb23758-logs\") pod \"182049e3-b96f-4c7e-b737-55375eb23758\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.142147 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-combined-ca-bundle\") pod \"182049e3-b96f-4c7e-b737-55375eb23758\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.142255 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-nova-metadata-tls-certs\") pod \"182049e3-b96f-4c7e-b737-55375eb23758\" (UID: \"182049e3-b96f-4c7e-b737-55375eb23758\") " Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.143425 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/182049e3-b96f-4c7e-b737-55375eb23758-logs" (OuterVolumeSpecName: "logs") pod "182049e3-b96f-4c7e-b737-55375eb23758" (UID: "182049e3-b96f-4c7e-b737-55375eb23758"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.151759 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/182049e3-b96f-4c7e-b737-55375eb23758-kube-api-access-n87q2" (OuterVolumeSpecName: "kube-api-access-n87q2") pod "182049e3-b96f-4c7e-b737-55375eb23758" (UID: "182049e3-b96f-4c7e-b737-55375eb23758"). InnerVolumeSpecName "kube-api-access-n87q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.225825 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "182049e3-b96f-4c7e-b737-55375eb23758" (UID: "182049e3-b96f-4c7e-b737-55375eb23758"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.320647 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.320675 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n87q2\" (UniqueName: \"kubernetes.io/projected/182049e3-b96f-4c7e-b737-55375eb23758-kube-api-access-n87q2\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.320687 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182049e3-b96f-4c7e-b737-55375eb23758-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.351720 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-config-data" (OuterVolumeSpecName: "config-data") pod "182049e3-b96f-4c7e-b737-55375eb23758" (UID: "182049e3-b96f-4c7e-b737-55375eb23758"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.354285 4741 generic.go:334] "Generic (PLEG): container finished" podID="182049e3-b96f-4c7e-b737-55375eb23758" containerID="cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067" exitCode=0 Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.354311 4741 generic.go:334] "Generic (PLEG): container finished" podID="182049e3-b96f-4c7e-b737-55375eb23758" containerID="ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad" exitCode=143 Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.354415 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"182049e3-b96f-4c7e-b737-55375eb23758","Type":"ContainerDied","Data":"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067"} Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.354442 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"182049e3-b96f-4c7e-b737-55375eb23758","Type":"ContainerDied","Data":"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad"} Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.354454 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"182049e3-b96f-4c7e-b737-55375eb23758","Type":"ContainerDied","Data":"dccb25610f36702dd33d7fdf6125bc01a80230de0e8dccb2db7485ad3aa170f5"} Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.354486 4741 scope.go:117] "RemoveContainer" containerID="cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.354593 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.384514 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "182049e3-b96f-4c7e-b737-55375eb23758" (UID: "182049e3-b96f-4c7e-b737-55375eb23758"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.392531 4741 generic.go:334] "Generic (PLEG): container finished" podID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerID="448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166" exitCode=143 Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.393267 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b082f7b-d8b9-4777-80c7-52f8f3b243cb","Type":"ContainerDied","Data":"448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166"} Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.398083 4741 scope.go:117] "RemoveContainer" containerID="ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.424316 4741 scope.go:117] "RemoveContainer" containerID="cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.424481 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.424516 4741 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/182049e3-b96f-4c7e-b737-55375eb23758-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:33 crc kubenswrapper[4741]: E0929 19:29:33.427743 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067\": container with ID starting with cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067 not found: ID does not exist" containerID="cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.427775 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067"} err="failed to get container status \"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067\": rpc error: code = NotFound desc = could not find container \"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067\": container with ID starting with cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067 not found: ID does not exist" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.427794 4741 scope.go:117] "RemoveContainer" containerID="ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad" Sep 29 19:29:33 crc kubenswrapper[4741]: E0929 19:29:33.428070 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad\": container with ID starting with ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad not found: ID does not exist" containerID="ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.428111 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad"} err="failed to get container status \"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad\": rpc error: code = NotFound desc = could not find container \"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad\": container with ID starting with ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad not found: ID does not exist" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.428126 4741 scope.go:117] "RemoveContainer" containerID="cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.428295 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067"} err="failed to get container status \"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067\": rpc error: code = NotFound desc = could not find container \"cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067\": container with ID starting with cb509bd6df4007c0c2d349c69100b9641fcd2ab8da84c6a6abd753245af3c067 not found: ID does not exist" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.428311 4741 scope.go:117] "RemoveContainer" containerID="ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.428310 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.428668 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad"} err="failed to get container status \"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad\": rpc error: code = NotFound desc = could not find container \"ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad\": container with ID starting with ef00196b802e0000ec87833678caabf2db5deb33c29f82bda7b902be71f84cad not found: ID does not exist" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.704124 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.715261 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.733382 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:33 crc kubenswrapper[4741]: E0929 19:29:33.733786 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-log" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.733803 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-log" Sep 29 19:29:33 crc kubenswrapper[4741]: E0929 19:29:33.733820 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-metadata" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.733827 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-metadata" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.734003 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-metadata" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.734019 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="182049e3-b96f-4c7e-b737-55375eb23758" containerName="nova-metadata-log" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.734963 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.737303 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.739066 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.753800 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.831493 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.831548 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzvqk\" (UniqueName: \"kubernetes.io/projected/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-kube-api-access-jzvqk\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.831575 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-config-data\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.831595 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.831661 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-logs\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.934075 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.934134 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzvqk\" (UniqueName: \"kubernetes.io/projected/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-kube-api-access-jzvqk\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.934186 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-config-data\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.935093 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.935214 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-logs\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.937162 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-logs\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.942609 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.953863 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-config-data\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.963491 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:33 crc kubenswrapper[4741]: I0929 19:29:33.971134 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzvqk\" (UniqueName: \"kubernetes.io/projected/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-kube-api-access-jzvqk\") pod \"nova-metadata-0\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " pod="openstack/nova-metadata-0" Sep 29 19:29:34 crc kubenswrapper[4741]: I0929 19:29:34.050527 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:29:34 crc kubenswrapper[4741]: I0929 19:29:34.409484 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93c5456d-7a59-4991-ad3d-58b04db78b24","Type":"ContainerStarted","Data":"aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070"} Sep 29 19:29:34 crc kubenswrapper[4741]: I0929 19:29:34.409927 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93c5456d-7a59-4991-ad3d-58b04db78b24","Type":"ContainerStarted","Data":"4c1e39c6e99eb3b23855e0ffbebc623383253f90eb1ac4a3ce9ccec667248510"} Sep 29 19:29:34 crc kubenswrapper[4741]: I0929 19:29:34.409595 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3ce9ebef-fbff-42be-8f52-3f949618ae3e" containerName="nova-scheduler-scheduler" containerID="cri-o://1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634" gracePeriod=30 Sep 29 19:29:34 crc kubenswrapper[4741]: I0929 19:29:34.433767 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.43375124 podStartE2EDuration="2.43375124s" podCreationTimestamp="2025-09-29 19:29:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:34.424922652 +0000 UTC m=+1216.072711984" watchObservedRunningTime="2025-09-29 19:29:34.43375124 +0000 UTC m=+1216.081540572" Sep 29 19:29:34 crc kubenswrapper[4741]: I0929 19:29:34.598594 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:29:34 crc kubenswrapper[4741]: W0929 19:29:34.608657 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cae00e0_9012_4f8b_96fd_ab24cd1b8ad4.slice/crio-c20bdc770ea5d8ae7d8bfbb83c8a9b7be3a18527b638c9b18c87373d0b652d2f WatchSource:0}: Error finding container c20bdc770ea5d8ae7d8bfbb83c8a9b7be3a18527b638c9b18c87373d0b652d2f: Status 404 returned error can't find the container with id c20bdc770ea5d8ae7d8bfbb83c8a9b7be3a18527b638c9b18c87373d0b652d2f Sep 29 19:29:35 crc kubenswrapper[4741]: I0929 19:29:35.113316 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="182049e3-b96f-4c7e-b737-55375eb23758" path="/var/lib/kubelet/pods/182049e3-b96f-4c7e-b737-55375eb23758/volumes" Sep 29 19:29:35 crc kubenswrapper[4741]: I0929 19:29:35.423564 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4","Type":"ContainerStarted","Data":"730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6"} Sep 29 19:29:35 crc kubenswrapper[4741]: I0929 19:29:35.423632 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4","Type":"ContainerStarted","Data":"8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1"} Sep 29 19:29:35 crc kubenswrapper[4741]: I0929 19:29:35.423648 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4","Type":"ContainerStarted","Data":"c20bdc770ea5d8ae7d8bfbb83c8a9b7be3a18527b638c9b18c87373d0b652d2f"} Sep 29 19:29:35 crc kubenswrapper[4741]: I0929 19:29:35.423682 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:35 crc kubenswrapper[4741]: I0929 19:29:35.448990 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.448967882 podStartE2EDuration="2.448967882s" podCreationTimestamp="2025-09-29 19:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:35.443019756 +0000 UTC m=+1217.090809088" watchObservedRunningTime="2025-09-29 19:29:35.448967882 +0000 UTC m=+1217.096757214" Sep 29 19:29:36 crc kubenswrapper[4741]: E0929 19:29:36.471053 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 19:29:36 crc kubenswrapper[4741]: E0929 19:29:36.480170 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 19:29:36 crc kubenswrapper[4741]: E0929 19:29:36.481927 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 19:29:36 crc kubenswrapper[4741]: E0929 19:29:36.481977 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3ce9ebef-fbff-42be-8f52-3f949618ae3e" containerName="nova-scheduler-scheduler" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.442004 4741 generic.go:334] "Generic (PLEG): container finished" podID="3ce9ebef-fbff-42be-8f52-3f949618ae3e" containerID="1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634" exitCode=0 Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.442090 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ce9ebef-fbff-42be-8f52-3f949618ae3e","Type":"ContainerDied","Data":"1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634"} Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.442328 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ce9ebef-fbff-42be-8f52-3f949618ae3e","Type":"ContainerDied","Data":"34cd3241853cbc39a8d19d2dafd5236fb51819d9960de42be53dcbca967879bb"} Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.442346 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34cd3241853cbc39a8d19d2dafd5236fb51819d9960de42be53dcbca967879bb" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.473851 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.603758 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-combined-ca-bundle\") pod \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.603842 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz7lx\" (UniqueName: \"kubernetes.io/projected/3ce9ebef-fbff-42be-8f52-3f949618ae3e-kube-api-access-fz7lx\") pod \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.603968 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-config-data\") pod \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\" (UID: \"3ce9ebef-fbff-42be-8f52-3f949618ae3e\") " Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.615058 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ce9ebef-fbff-42be-8f52-3f949618ae3e-kube-api-access-fz7lx" (OuterVolumeSpecName: "kube-api-access-fz7lx") pod "3ce9ebef-fbff-42be-8f52-3f949618ae3e" (UID: "3ce9ebef-fbff-42be-8f52-3f949618ae3e"). InnerVolumeSpecName "kube-api-access-fz7lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.635573 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ce9ebef-fbff-42be-8f52-3f949618ae3e" (UID: "3ce9ebef-fbff-42be-8f52-3f949618ae3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.639599 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-config-data" (OuterVolumeSpecName: "config-data") pod "3ce9ebef-fbff-42be-8f52-3f949618ae3e" (UID: "3ce9ebef-fbff-42be-8f52-3f949618ae3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.705698 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.705757 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz7lx\" (UniqueName: \"kubernetes.io/projected/3ce9ebef-fbff-42be-8f52-3f949618ae3e-kube-api-access-fz7lx\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:37 crc kubenswrapper[4741]: I0929 19:29:37.705772 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce9ebef-fbff-42be-8f52-3f949618ae3e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.338548 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.432786 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-combined-ca-bundle\") pod \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.433190 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-logs\") pod \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.433243 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbm9x\" (UniqueName: \"kubernetes.io/projected/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-kube-api-access-gbm9x\") pod \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.433332 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-config-data\") pod \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\" (UID: \"5b082f7b-d8b9-4777-80c7-52f8f3b243cb\") " Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.434197 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-logs" (OuterVolumeSpecName: "logs") pod "5b082f7b-d8b9-4777-80c7-52f8f3b243cb" (UID: "5b082f7b-d8b9-4777-80c7-52f8f3b243cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.437606 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-kube-api-access-gbm9x" (OuterVolumeSpecName: "kube-api-access-gbm9x") pod "5b082f7b-d8b9-4777-80c7-52f8f3b243cb" (UID: "5b082f7b-d8b9-4777-80c7-52f8f3b243cb"). InnerVolumeSpecName "kube-api-access-gbm9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.453590 4741 generic.go:334] "Generic (PLEG): container finished" podID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerID="8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a" exitCode=0 Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.453672 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b082f7b-d8b9-4777-80c7-52f8f3b243cb","Type":"ContainerDied","Data":"8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a"} Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.453717 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b082f7b-d8b9-4777-80c7-52f8f3b243cb","Type":"ContainerDied","Data":"739413966ca129e79a20a040545df098e42f8153f0d4216cd2fbdc624b6dd9ef"} Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.453721 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.453731 4741 scope.go:117] "RemoveContainer" containerID="8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.454048 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.459194 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b082f7b-d8b9-4777-80c7-52f8f3b243cb" (UID: "5b082f7b-d8b9-4777-80c7-52f8f3b243cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.462645 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-config-data" (OuterVolumeSpecName: "config-data") pod "5b082f7b-d8b9-4777-80c7-52f8f3b243cb" (UID: "5b082f7b-d8b9-4777-80c7-52f8f3b243cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.534938 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.534966 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.534976 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbm9x\" (UniqueName: \"kubernetes.io/projected/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-kube-api-access-gbm9x\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.534988 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b082f7b-d8b9-4777-80c7-52f8f3b243cb-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.542525 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.547113 4741 scope.go:117] "RemoveContainer" containerID="448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.553810 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.562137 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: E0929 19:29:38.562606 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-log" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.562626 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-log" Sep 29 19:29:38 crc kubenswrapper[4741]: E0929 19:29:38.562659 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ce9ebef-fbff-42be-8f52-3f949618ae3e" containerName="nova-scheduler-scheduler" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.562671 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ce9ebef-fbff-42be-8f52-3f949618ae3e" containerName="nova-scheduler-scheduler" Sep 29 19:29:38 crc kubenswrapper[4741]: E0929 19:29:38.562685 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-api" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.562693 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-api" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.562921 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-log" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.562937 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" containerName="nova-api-api" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.562967 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ce9ebef-fbff-42be-8f52-3f949618ae3e" containerName="nova-scheduler-scheduler" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.563693 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.571154 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.576047 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.586606 4741 scope.go:117] "RemoveContainer" containerID="8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a" Sep 29 19:29:38 crc kubenswrapper[4741]: E0929 19:29:38.587130 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a\": container with ID starting with 8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a not found: ID does not exist" containerID="8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.587161 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a"} err="failed to get container status \"8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a\": rpc error: code = NotFound desc = could not find container \"8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a\": container with ID starting with 8dcc0df57eb6d94f19fe9fb6175988d6d1a2165552aab88ce3d330138fb7e75a not found: ID does not exist" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.587182 4741 scope.go:117] "RemoveContainer" containerID="448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166" Sep 29 19:29:38 crc kubenswrapper[4741]: E0929 19:29:38.587612 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166\": container with ID starting with 448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166 not found: ID does not exist" containerID="448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.587636 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166"} err="failed to get container status \"448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166\": rpc error: code = NotFound desc = could not find container \"448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166\": container with ID starting with 448a332c301e837161c9d2d500e5e7c427de2f7ba083530f7c99a863c904e166 not found: ID does not exist" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.636374 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlvj5\" (UniqueName: \"kubernetes.io/projected/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-kube-api-access-nlvj5\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.636631 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.636793 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-config-data\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.738032 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.738083 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-config-data\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.738128 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlvj5\" (UniqueName: \"kubernetes.io/projected/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-kube-api-access-nlvj5\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.743019 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-config-data\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.743228 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.758115 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlvj5\" (UniqueName: \"kubernetes.io/projected/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-kube-api-access-nlvj5\") pod \"nova-scheduler-0\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.796331 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.804547 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.814732 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.816303 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.819044 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.823012 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.839354 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.839469 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1e31bb-969a-4f4b-9826-9b77df8c244b-logs\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.839510 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsc7n\" (UniqueName: \"kubernetes.io/projected/2f1e31bb-969a-4f4b-9826-9b77df8c244b-kube-api-access-jsc7n\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.839538 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-config-data\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.895700 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.940805 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1e31bb-969a-4f4b-9826-9b77df8c244b-logs\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.940862 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsc7n\" (UniqueName: \"kubernetes.io/projected/2f1e31bb-969a-4f4b-9826-9b77df8c244b-kube-api-access-jsc7n\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.940930 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-config-data\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.941090 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.941940 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1e31bb-969a-4f4b-9826-9b77df8c244b-logs\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.947021 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.948495 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-config-data\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:38 crc kubenswrapper[4741]: I0929 19:29:38.969934 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsc7n\" (UniqueName: \"kubernetes.io/projected/2f1e31bb-969a-4f4b-9826-9b77df8c244b-kube-api-access-jsc7n\") pod \"nova-api-0\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " pod="openstack/nova-api-0" Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.051382 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.051494 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.097356 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ce9ebef-fbff-42be-8f52-3f949618ae3e" path="/var/lib/kubelet/pods/3ce9ebef-fbff-42be-8f52-3f949618ae3e/volumes" Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.097961 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b082f7b-d8b9-4777-80c7-52f8f3b243cb" path="/var/lib/kubelet/pods/5b082f7b-d8b9-4777-80c7-52f8f3b243cb/volumes" Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.138245 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.301564 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:29:39 crc kubenswrapper[4741]: W0929 19:29:39.309548 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc0d61fc_e880_4857_bdea_b90f2a86b4c2.slice/crio-9cd20a1e71ce225dec4b19917b45293e7176f419aad8b03b7a49b4339d607779 WatchSource:0}: Error finding container 9cd20a1e71ce225dec4b19917b45293e7176f419aad8b03b7a49b4339d607779: Status 404 returned error can't find the container with id 9cd20a1e71ce225dec4b19917b45293e7176f419aad8b03b7a49b4339d607779 Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.464574 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc0d61fc-e880-4857-bdea-b90f2a86b4c2","Type":"ContainerStarted","Data":"9cd20a1e71ce225dec4b19917b45293e7176f419aad8b03b7a49b4339d607779"} Sep 29 19:29:39 crc kubenswrapper[4741]: I0929 19:29:39.559164 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:29:39 crc kubenswrapper[4741]: W0929 19:29:39.563231 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f1e31bb_969a_4f4b_9826_9b77df8c244b.slice/crio-d9e4660888a9955b1cbbd75b284d20d2ca3a721155b7508dde852eef26bec17a WatchSource:0}: Error finding container d9e4660888a9955b1cbbd75b284d20d2ca3a721155b7508dde852eef26bec17a: Status 404 returned error can't find the container with id d9e4660888a9955b1cbbd75b284d20d2ca3a721155b7508dde852eef26bec17a Sep 29 19:29:40 crc kubenswrapper[4741]: I0929 19:29:40.477101 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f1e31bb-969a-4f4b-9826-9b77df8c244b","Type":"ContainerStarted","Data":"0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf"} Sep 29 19:29:40 crc kubenswrapper[4741]: I0929 19:29:40.477450 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f1e31bb-969a-4f4b-9826-9b77df8c244b","Type":"ContainerStarted","Data":"d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff"} Sep 29 19:29:40 crc kubenswrapper[4741]: I0929 19:29:40.477467 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f1e31bb-969a-4f4b-9826-9b77df8c244b","Type":"ContainerStarted","Data":"d9e4660888a9955b1cbbd75b284d20d2ca3a721155b7508dde852eef26bec17a"} Sep 29 19:29:40 crc kubenswrapper[4741]: I0929 19:29:40.478983 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc0d61fc-e880-4857-bdea-b90f2a86b4c2","Type":"ContainerStarted","Data":"7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4"} Sep 29 19:29:40 crc kubenswrapper[4741]: I0929 19:29:40.510968 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.510949259 podStartE2EDuration="2.510949259s" podCreationTimestamp="2025-09-29 19:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:40.505866819 +0000 UTC m=+1222.153656151" watchObservedRunningTime="2025-09-29 19:29:40.510949259 +0000 UTC m=+1222.158738591" Sep 29 19:29:40 crc kubenswrapper[4741]: I0929 19:29:40.526706 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.526690874 podStartE2EDuration="2.526690874s" podCreationTimestamp="2025-09-29 19:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:40.520428827 +0000 UTC m=+1222.168218169" watchObservedRunningTime="2025-09-29 19:29:40.526690874 +0000 UTC m=+1222.174480206" Sep 29 19:29:42 crc kubenswrapper[4741]: I0929 19:29:42.783860 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 29 19:29:43 crc kubenswrapper[4741]: I0929 19:29:43.896071 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 19:29:44 crc kubenswrapper[4741]: I0929 19:29:44.051588 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 19:29:44 crc kubenswrapper[4741]: I0929 19:29:44.051880 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 19:29:45 crc kubenswrapper[4741]: I0929 19:29:45.070638 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 19:29:45 crc kubenswrapper[4741]: I0929 19:29:45.070708 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 19:29:46 crc kubenswrapper[4741]: I0929 19:29:46.340280 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 29 19:29:48 crc kubenswrapper[4741]: I0929 19:29:48.896563 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 19:29:48 crc kubenswrapper[4741]: I0929 19:29:48.927049 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 19:29:49 crc kubenswrapper[4741]: I0929 19:29:49.139366 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 19:29:49 crc kubenswrapper[4741]: I0929 19:29:49.139428 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 19:29:49 crc kubenswrapper[4741]: I0929 19:29:49.624503 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:29:49 crc kubenswrapper[4741]: I0929 19:29:49.625327 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="84df25f5-0db8-457f-b964-3cdb332ebf02" containerName="kube-state-metrics" containerID="cri-o://69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000" gracePeriod=30 Sep 29 19:29:49 crc kubenswrapper[4741]: I0929 19:29:49.655742 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.128554 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.168275 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6xk8\" (UniqueName: \"kubernetes.io/projected/84df25f5-0db8-457f-b964-3cdb332ebf02-kube-api-access-r6xk8\") pod \"84df25f5-0db8-457f-b964-3cdb332ebf02\" (UID: \"84df25f5-0db8-457f-b964-3cdb332ebf02\") " Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.179790 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84df25f5-0db8-457f-b964-3cdb332ebf02-kube-api-access-r6xk8" (OuterVolumeSpecName: "kube-api-access-r6xk8") pod "84df25f5-0db8-457f-b964-3cdb332ebf02" (UID: "84df25f5-0db8-457f-b964-3cdb332ebf02"). InnerVolumeSpecName "kube-api-access-r6xk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.223613 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.223608 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.270327 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6xk8\" (UniqueName: \"kubernetes.io/projected/84df25f5-0db8-457f-b964-3cdb332ebf02-kube-api-access-r6xk8\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.583030 4741 generic.go:334] "Generic (PLEG): container finished" podID="84df25f5-0db8-457f-b964-3cdb332ebf02" containerID="69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000" exitCode=2 Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.583075 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84df25f5-0db8-457f-b964-3cdb332ebf02","Type":"ContainerDied","Data":"69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000"} Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.583374 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84df25f5-0db8-457f-b964-3cdb332ebf02","Type":"ContainerDied","Data":"e4eb82a5251b0f6c6ed3f995f8fe5db16c66735b092c8672e8f311e3a9850700"} Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.583428 4741 scope.go:117] "RemoveContainer" containerID="69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.583123 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.606369 4741 scope.go:117] "RemoveContainer" containerID="69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000" Sep 29 19:29:50 crc kubenswrapper[4741]: E0929 19:29:50.607428 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000\": container with ID starting with 69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000 not found: ID does not exist" containerID="69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.607471 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000"} err="failed to get container status \"69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000\": rpc error: code = NotFound desc = could not find container \"69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000\": container with ID starting with 69cd1bdecf9ddd5ba1326a7f26d0a7f6d1924449fe3796bde414f69104ad3000 not found: ID does not exist" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.625383 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.636370 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.653213 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:29:50 crc kubenswrapper[4741]: E0929 19:29:50.653773 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84df25f5-0db8-457f-b964-3cdb332ebf02" containerName="kube-state-metrics" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.653796 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84df25f5-0db8-457f-b964-3cdb332ebf02" containerName="kube-state-metrics" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.654012 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="84df25f5-0db8-457f-b964-3cdb332ebf02" containerName="kube-state-metrics" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.654683 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.657551 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.657864 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.676687 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbk2f\" (UniqueName: \"kubernetes.io/projected/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-api-access-wbk2f\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.676761 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.676834 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.676888 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.682823 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.778686 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbk2f\" (UniqueName: \"kubernetes.io/projected/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-api-access-wbk2f\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.778753 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.778800 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.778824 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.786485 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.786856 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.799449 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbk2f\" (UniqueName: \"kubernetes.io/projected/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-api-access-wbk2f\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.804927 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " pod="openstack/kube-state-metrics-0" Sep 29 19:29:50 crc kubenswrapper[4741]: I0929 19:29:50.978069 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.100562 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84df25f5-0db8-457f-b964-3cdb332ebf02" path="/var/lib/kubelet/pods/84df25f5-0db8-457f-b964-3cdb332ebf02/volumes" Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.427889 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:29:51 crc kubenswrapper[4741]: W0929 19:29:51.435728 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c39883d_b9e0_4b1f_a7b0_8d29c04d066a.slice/crio-bdb6ca44925061fd0dc3065188cdf0b6fadd4d39e0bf0c0bf750a7c29bcc594c WatchSource:0}: Error finding container bdb6ca44925061fd0dc3065188cdf0b6fadd4d39e0bf0c0bf750a7c29bcc594c: Status 404 returned error can't find the container with id bdb6ca44925061fd0dc3065188cdf0b6fadd4d39e0bf0c0bf750a7c29bcc594c Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.618053 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a","Type":"ContainerStarted","Data":"bdb6ca44925061fd0dc3065188cdf0b6fadd4d39e0bf0c0bf750a7c29bcc594c"} Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.625073 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.625555 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-central-agent" containerID="cri-o://5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d" gracePeriod=30 Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.625812 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="proxy-httpd" containerID="cri-o://c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891" gracePeriod=30 Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.625962 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="sg-core" containerID="cri-o://5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9" gracePeriod=30 Sep 29 19:29:51 crc kubenswrapper[4741]: I0929 19:29:51.626000 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-notification-agent" containerID="cri-o://b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef" gracePeriod=30 Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.629212 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a","Type":"ContainerStarted","Data":"7720681edbf8321fa29bc81ceddd38f098ef1fa29db00e81fd9806812c014bd7"} Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.629535 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.632428 4741 generic.go:334] "Generic (PLEG): container finished" podID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerID="c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891" exitCode=0 Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.632452 4741 generic.go:334] "Generic (PLEG): container finished" podID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerID="5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9" exitCode=2 Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.632460 4741 generic.go:334] "Generic (PLEG): container finished" podID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerID="5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d" exitCode=0 Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.632502 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerDied","Data":"c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891"} Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.632546 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerDied","Data":"5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9"} Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.632558 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerDied","Data":"5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d"} Sep 29 19:29:52 crc kubenswrapper[4741]: I0929 19:29:52.651853 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.175150462 podStartE2EDuration="2.651823261s" podCreationTimestamp="2025-09-29 19:29:50 +0000 UTC" firstStartedPulling="2025-09-29 19:29:51.438106767 +0000 UTC m=+1233.085896109" lastFinishedPulling="2025-09-29 19:29:51.914779576 +0000 UTC m=+1233.562568908" observedRunningTime="2025-09-29 19:29:52.643704845 +0000 UTC m=+1234.291494177" watchObservedRunningTime="2025-09-29 19:29:52.651823261 +0000 UTC m=+1234.299612593" Sep 29 19:29:54 crc kubenswrapper[4741]: I0929 19:29:54.057162 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 19:29:54 crc kubenswrapper[4741]: I0929 19:29:54.057590 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 19:29:54 crc kubenswrapper[4741]: I0929 19:29:54.067558 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 19:29:54 crc kubenswrapper[4741]: I0929 19:29:54.070061 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.123269 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.256647 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-run-httpd\") pod \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.256702 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-log-httpd\") pod \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.257097 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "760f134b-0672-4cf2-a4c7-0d6ffea1db4a" (UID: "760f134b-0672-4cf2-a4c7-0d6ffea1db4a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.257239 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "760f134b-0672-4cf2-a4c7-0d6ffea1db4a" (UID: "760f134b-0672-4cf2-a4c7-0d6ffea1db4a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.257298 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-config-data\") pod \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.257382 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-sg-core-conf-yaml\") pod \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.257709 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-combined-ca-bundle\") pod \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.257751 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrxrr\" (UniqueName: \"kubernetes.io/projected/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-kube-api-access-zrxrr\") pod \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.257813 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-scripts\") pod \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\" (UID: \"760f134b-0672-4cf2-a4c7-0d6ffea1db4a\") " Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.258107 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.258127 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.262512 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-kube-api-access-zrxrr" (OuterVolumeSpecName: "kube-api-access-zrxrr") pod "760f134b-0672-4cf2-a4c7-0d6ffea1db4a" (UID: "760f134b-0672-4cf2-a4c7-0d6ffea1db4a"). InnerVolumeSpecName "kube-api-access-zrxrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.263484 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-scripts" (OuterVolumeSpecName: "scripts") pod "760f134b-0672-4cf2-a4c7-0d6ffea1db4a" (UID: "760f134b-0672-4cf2-a4c7-0d6ffea1db4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.285761 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "760f134b-0672-4cf2-a4c7-0d6ffea1db4a" (UID: "760f134b-0672-4cf2-a4c7-0d6ffea1db4a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.330726 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "760f134b-0672-4cf2-a4c7-0d6ffea1db4a" (UID: "760f134b-0672-4cf2-a4c7-0d6ffea1db4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.347824 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-config-data" (OuterVolumeSpecName: "config-data") pod "760f134b-0672-4cf2-a4c7-0d6ffea1db4a" (UID: "760f134b-0672-4cf2-a4c7-0d6ffea1db4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.359496 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.359525 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrxrr\" (UniqueName: \"kubernetes.io/projected/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-kube-api-access-zrxrr\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.359537 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.359545 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.359553 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/760f134b-0672-4cf2-a4c7-0d6ffea1db4a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.663648 4741 generic.go:334] "Generic (PLEG): container finished" podID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerID="b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef" exitCode=0 Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.663721 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.663742 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerDied","Data":"b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef"} Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.664094 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"760f134b-0672-4cf2-a4c7-0d6ffea1db4a","Type":"ContainerDied","Data":"6f551232eb1334d309414df32f45f281406c2fb233cc52efd4db1fad53115af6"} Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.664120 4741 scope.go:117] "RemoveContainer" containerID="c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.694623 4741 scope.go:117] "RemoveContainer" containerID="5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.699465 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.708426 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.723423 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.724272 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-central-agent" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724296 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-central-agent" Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.724319 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-notification-agent" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724330 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-notification-agent" Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.724361 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="proxy-httpd" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724369 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="proxy-httpd" Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.724405 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="sg-core" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724413 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="sg-core" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724625 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="sg-core" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724647 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-notification-agent" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724658 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="proxy-httpd" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.724679 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" containerName="ceilometer-central-agent" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.727012 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.731036 4741 scope.go:117] "RemoveContainer" containerID="b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.732539 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.733005 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.733272 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.770070 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.810884 4741 scope.go:117] "RemoveContainer" containerID="5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.838083 4741 scope.go:117] "RemoveContainer" containerID="c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891" Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.838732 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891\": container with ID starting with c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891 not found: ID does not exist" containerID="c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.838801 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891"} err="failed to get container status \"c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891\": rpc error: code = NotFound desc = could not find container \"c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891\": container with ID starting with c06527929a20db9f4557393dab716f7c58747cf44464e9dfb9b864fd294ad891 not found: ID does not exist" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.838876 4741 scope.go:117] "RemoveContainer" containerID="5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9" Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.839221 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9\": container with ID starting with 5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9 not found: ID does not exist" containerID="5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.839283 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9"} err="failed to get container status \"5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9\": rpc error: code = NotFound desc = could not find container \"5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9\": container with ID starting with 5e8b5d0f8c95399948a2ad55cdf28be6c99987177e90408d9404d241aedbbeb9 not found: ID does not exist" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.839326 4741 scope.go:117] "RemoveContainer" containerID="b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef" Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.839898 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef\": container with ID starting with b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef not found: ID does not exist" containerID="b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.839926 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef"} err="failed to get container status \"b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef\": rpc error: code = NotFound desc = could not find container \"b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef\": container with ID starting with b9568ada24cc2ed7b61ca78cc7e711f4610ee471f3c442c9d3fd3a056b7ab0ef not found: ID does not exist" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.839965 4741 scope.go:117] "RemoveContainer" containerID="5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d" Sep 29 19:29:55 crc kubenswrapper[4741]: E0929 19:29:55.840186 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d\": container with ID starting with 5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d not found: ID does not exist" containerID="5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.840207 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d"} err="failed to get container status \"5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d\": rpc error: code = NotFound desc = could not find container \"5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d\": container with ID starting with 5c8827b75351e12cce83be66c8dd14b9ae97376996365e32ae1723248bafd55d not found: ID does not exist" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.869679 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-log-httpd\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.869724 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.869748 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.869778 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2llvp\" (UniqueName: \"kubernetes.io/projected/84280a33-f064-48dc-a074-6283a759cacf-kube-api-access-2llvp\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.869795 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-config-data\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.869819 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.870073 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-run-httpd\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.870142 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-scripts\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.971793 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2llvp\" (UniqueName: \"kubernetes.io/projected/84280a33-f064-48dc-a074-6283a759cacf-kube-api-access-2llvp\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972127 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-config-data\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972160 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972197 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-run-httpd\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972221 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-scripts\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972379 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-log-httpd\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972420 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972449 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972790 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-log-httpd\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.972957 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-run-httpd\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.976476 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.976606 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-scripts\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.976636 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.977341 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-config-data\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.987563 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:55 crc kubenswrapper[4741]: I0929 19:29:55.988010 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2llvp\" (UniqueName: \"kubernetes.io/projected/84280a33-f064-48dc-a074-6283a759cacf-kube-api-access-2llvp\") pod \"ceilometer-0\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " pod="openstack/ceilometer-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.100547 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.477233 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.486264 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-config-data\") pod \"c152f034-8b8e-44cf-8b05-48da43cd9378\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.486634 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-combined-ca-bundle\") pod \"c152f034-8b8e-44cf-8b05-48da43cd9378\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.486728 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn44d\" (UniqueName: \"kubernetes.io/projected/c152f034-8b8e-44cf-8b05-48da43cd9378-kube-api-access-mn44d\") pod \"c152f034-8b8e-44cf-8b05-48da43cd9378\" (UID: \"c152f034-8b8e-44cf-8b05-48da43cd9378\") " Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.526971 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c152f034-8b8e-44cf-8b05-48da43cd9378-kube-api-access-mn44d" (OuterVolumeSpecName: "kube-api-access-mn44d") pod "c152f034-8b8e-44cf-8b05-48da43cd9378" (UID: "c152f034-8b8e-44cf-8b05-48da43cd9378"). InnerVolumeSpecName "kube-api-access-mn44d". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.531049 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c152f034-8b8e-44cf-8b05-48da43cd9378" (UID: "c152f034-8b8e-44cf-8b05-48da43cd9378"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.531606 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-config-data" (OuterVolumeSpecName: "config-data") pod "c152f034-8b8e-44cf-8b05-48da43cd9378" (UID: "c152f034-8b8e-44cf-8b05-48da43cd9378"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.567426 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:29:56 crc kubenswrapper[4741]: W0929 19:29:56.572970 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84280a33_f064_48dc_a074_6283a759cacf.slice/crio-7d0f03bd6f71033fae46dc988d44acc17ad8cc309e903194d3a97de50dd9ccee WatchSource:0}: Error finding container 7d0f03bd6f71033fae46dc988d44acc17ad8cc309e903194d3a97de50dd9ccee: Status 404 returned error can't find the container with id 7d0f03bd6f71033fae46dc988d44acc17ad8cc309e903194d3a97de50dd9ccee Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.589777 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.589829 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn44d\" (UniqueName: \"kubernetes.io/projected/c152f034-8b8e-44cf-8b05-48da43cd9378-kube-api-access-mn44d\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.589847 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c152f034-8b8e-44cf-8b05-48da43cd9378-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.673269 4741 generic.go:334] "Generic (PLEG): container finished" podID="c152f034-8b8e-44cf-8b05-48da43cd9378" containerID="4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2" exitCode=137 Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.673428 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.673360 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c152f034-8b8e-44cf-8b05-48da43cd9378","Type":"ContainerDied","Data":"4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2"} Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.673489 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c152f034-8b8e-44cf-8b05-48da43cd9378","Type":"ContainerDied","Data":"fc1ed0ee09a7879f8c1c6a7b9f42420e032e2ba552f050c758e63eab83a92fd3"} Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.673514 4741 scope.go:117] "RemoveContainer" containerID="4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.675714 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerStarted","Data":"7d0f03bd6f71033fae46dc988d44acc17ad8cc309e903194d3a97de50dd9ccee"} Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.706631 4741 scope.go:117] "RemoveContainer" containerID="4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2" Sep 29 19:29:56 crc kubenswrapper[4741]: E0929 19:29:56.714490 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2\": container with ID starting with 4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2 not found: ID does not exist" containerID="4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.714720 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2"} err="failed to get container status \"4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2\": rpc error: code = NotFound desc = could not find container \"4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2\": container with ID starting with 4f1fd9a52849a42b9d1a4ead832b893ed3c14ab7fb8349c43cbcdafd80559ad2 not found: ID does not exist" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.722560 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.741665 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.752832 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:56 crc kubenswrapper[4741]: E0929 19:29:56.753351 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c152f034-8b8e-44cf-8b05-48da43cd9378" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.753372 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c152f034-8b8e-44cf-8b05-48da43cd9378" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.753632 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c152f034-8b8e-44cf-8b05-48da43cd9378" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.754507 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.760323 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.760529 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.764623 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.775046 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.798025 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.798237 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.798302 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.798474 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rvch\" (UniqueName: \"kubernetes.io/projected/f24ec6be-6832-4eee-916c-2237e81bf0f1-kube-api-access-9rvch\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.798573 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.899862 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.900218 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.900932 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.901014 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rvch\" (UniqueName: \"kubernetes.io/projected/f24ec6be-6832-4eee-916c-2237e81bf0f1-kube-api-access-9rvch\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.901054 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.903484 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.903782 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.905099 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.905122 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:56 crc kubenswrapper[4741]: I0929 19:29:56.928228 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rvch\" (UniqueName: \"kubernetes.io/projected/f24ec6be-6832-4eee-916c-2237e81bf0f1-kube-api-access-9rvch\") pod \"nova-cell1-novncproxy-0\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:57 crc kubenswrapper[4741]: I0929 19:29:57.096109 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="760f134b-0672-4cf2-a4c7-0d6ffea1db4a" path="/var/lib/kubelet/pods/760f134b-0672-4cf2-a4c7-0d6ffea1db4a/volumes" Sep 29 19:29:57 crc kubenswrapper[4741]: I0929 19:29:57.096961 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c152f034-8b8e-44cf-8b05-48da43cd9378" path="/var/lib/kubelet/pods/c152f034-8b8e-44cf-8b05-48da43cd9378/volumes" Sep 29 19:29:57 crc kubenswrapper[4741]: I0929 19:29:57.114071 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:29:57 crc kubenswrapper[4741]: I0929 19:29:57.573111 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:29:57 crc kubenswrapper[4741]: W0929 19:29:57.577545 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf24ec6be_6832_4eee_916c_2237e81bf0f1.slice/crio-78582c14d813d4691bde7fb531ce55242e16355d419af06008602af4258d247a WatchSource:0}: Error finding container 78582c14d813d4691bde7fb531ce55242e16355d419af06008602af4258d247a: Status 404 returned error can't find the container with id 78582c14d813d4691bde7fb531ce55242e16355d419af06008602af4258d247a Sep 29 19:29:57 crc kubenswrapper[4741]: I0929 19:29:57.688046 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerStarted","Data":"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf"} Sep 29 19:29:57 crc kubenswrapper[4741]: I0929 19:29:57.689484 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f24ec6be-6832-4eee-916c-2237e81bf0f1","Type":"ContainerStarted","Data":"78582c14d813d4691bde7fb531ce55242e16355d419af06008602af4258d247a"} Sep 29 19:29:58 crc kubenswrapper[4741]: I0929 19:29:58.703197 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerStarted","Data":"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285"} Sep 29 19:29:58 crc kubenswrapper[4741]: I0929 19:29:58.705608 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f24ec6be-6832-4eee-916c-2237e81bf0f1","Type":"ContainerStarted","Data":"e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3"} Sep 29 19:29:58 crc kubenswrapper[4741]: I0929 19:29:58.736085 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.736059261 podStartE2EDuration="2.736059261s" podCreationTimestamp="2025-09-29 19:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:29:58.727179922 +0000 UTC m=+1240.374969294" watchObservedRunningTime="2025-09-29 19:29:58.736059261 +0000 UTC m=+1240.383848673" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.143206 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.143948 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.147448 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.147716 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.724622 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerStarted","Data":"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26"} Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.725072 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.729438 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.902739 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-55hm7"] Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.904447 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.933460 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-55hm7"] Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.976631 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-config\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.976706 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.976735 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8jc4\" (UniqueName: \"kubernetes.io/projected/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-kube-api-access-f8jc4\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.976782 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.976871 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:29:59 crc kubenswrapper[4741]: I0929 19:29:59.976893 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.078104 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.078147 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8jc4\" (UniqueName: \"kubernetes.io/projected/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-kube-api-access-f8jc4\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.078187 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.078271 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.078291 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.078313 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-config\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.079303 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.079569 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.079617 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.079760 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-config\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.080108 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.098214 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8jc4\" (UniqueName: \"kubernetes.io/projected/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-kube-api-access-f8jc4\") pod \"dnsmasq-dns-89c5cd4d5-55hm7\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.135145 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595"] Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.136624 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.138517 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.138887 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.146349 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595"] Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.181464 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22434baf-68e0-4783-8835-27fa1e05b00d-secret-volume\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.181521 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22434baf-68e0-4783-8835-27fa1e05b00d-config-volume\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.181563 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxn7p\" (UniqueName: \"kubernetes.io/projected/22434baf-68e0-4783-8835-27fa1e05b00d-kube-api-access-vxn7p\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.235653 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.283324 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22434baf-68e0-4783-8835-27fa1e05b00d-config-volume\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.283383 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxn7p\" (UniqueName: \"kubernetes.io/projected/22434baf-68e0-4783-8835-27fa1e05b00d-kube-api-access-vxn7p\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.283599 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22434baf-68e0-4783-8835-27fa1e05b00d-secret-volume\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.284285 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22434baf-68e0-4783-8835-27fa1e05b00d-config-volume\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.287358 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22434baf-68e0-4783-8835-27fa1e05b00d-secret-volume\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.302153 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxn7p\" (UniqueName: \"kubernetes.io/projected/22434baf-68e0-4783-8835-27fa1e05b00d-kube-api-access-vxn7p\") pod \"collect-profiles-29319570-lt595\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.491622 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.779317 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-55hm7"] Sep 29 19:30:00 crc kubenswrapper[4741]: I0929 19:30:00.998784 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.003952 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595"] Sep 29 19:30:01 crc kubenswrapper[4741]: W0929 19:30:01.005812 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22434baf_68e0_4783_8835_27fa1e05b00d.slice/crio-818a96eaffb30c26e30541e62b1c2ba1eeff9ae84cc7c97a1e7f2f25dff6485d WatchSource:0}: Error finding container 818a96eaffb30c26e30541e62b1c2ba1eeff9ae84cc7c97a1e7f2f25dff6485d: Status 404 returned error can't find the container with id 818a96eaffb30c26e30541e62b1c2ba1eeff9ae84cc7c97a1e7f2f25dff6485d Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.762917 4741 generic.go:334] "Generic (PLEG): container finished" podID="22434baf-68e0-4783-8835-27fa1e05b00d" containerID="6d5598999cc4956e0cf131e4f176e34e0fe22baa68c7086cad3d29abcdeb7336" exitCode=0 Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.763010 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" event={"ID":"22434baf-68e0-4783-8835-27fa1e05b00d","Type":"ContainerDied","Data":"6d5598999cc4956e0cf131e4f176e34e0fe22baa68c7086cad3d29abcdeb7336"} Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.763297 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" event={"ID":"22434baf-68e0-4783-8835-27fa1e05b00d","Type":"ContainerStarted","Data":"818a96eaffb30c26e30541e62b1c2ba1eeff9ae84cc7c97a1e7f2f25dff6485d"} Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.767323 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerStarted","Data":"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94"} Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.767384 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.771540 4741 generic.go:334] "Generic (PLEG): container finished" podID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerID="056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d" exitCode=0 Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.771630 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" event={"ID":"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb","Type":"ContainerDied","Data":"056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d"} Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.771652 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" event={"ID":"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb","Type":"ContainerStarted","Data":"afe473e033f4be1934d8ed1c53f5d257398a66f8ee46f1776279b0e6cc62bd9c"} Sep 29 19:30:01 crc kubenswrapper[4741]: I0929 19:30:01.807370 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7742912 podStartE2EDuration="6.807352893s" podCreationTimestamp="2025-09-29 19:29:55 +0000 UTC" firstStartedPulling="2025-09-29 19:29:56.575799085 +0000 UTC m=+1238.223588417" lastFinishedPulling="2025-09-29 19:30:00.608860778 +0000 UTC m=+1242.256650110" observedRunningTime="2025-09-29 19:30:01.803917845 +0000 UTC m=+1243.451707197" watchObservedRunningTime="2025-09-29 19:30:01.807352893 +0000 UTC m=+1243.455142225" Sep 29 19:30:02 crc kubenswrapper[4741]: I0929 19:30:02.115086 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:30:02 crc kubenswrapper[4741]: I0929 19:30:02.187357 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:02 crc kubenswrapper[4741]: I0929 19:30:02.782969 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" event={"ID":"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb","Type":"ContainerStarted","Data":"1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da"} Sep 29 19:30:02 crc kubenswrapper[4741]: I0929 19:30:02.783623 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-log" containerID="cri-o://d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff" gracePeriod=30 Sep 29 19:30:02 crc kubenswrapper[4741]: I0929 19:30:02.783727 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-api" containerID="cri-o://0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf" gracePeriod=30 Sep 29 19:30:02 crc kubenswrapper[4741]: I0929 19:30:02.821123 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" podStartSLOduration=3.8211019090000002 podStartE2EDuration="3.821101909s" podCreationTimestamp="2025-09-29 19:29:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:30:02.809602448 +0000 UTC m=+1244.457391800" watchObservedRunningTime="2025-09-29 19:30:02.821101909 +0000 UTC m=+1244.468891241" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.074500 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.164293 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.341354 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxn7p\" (UniqueName: \"kubernetes.io/projected/22434baf-68e0-4783-8835-27fa1e05b00d-kube-api-access-vxn7p\") pod \"22434baf-68e0-4783-8835-27fa1e05b00d\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.341476 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22434baf-68e0-4783-8835-27fa1e05b00d-secret-volume\") pod \"22434baf-68e0-4783-8835-27fa1e05b00d\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.341548 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22434baf-68e0-4783-8835-27fa1e05b00d-config-volume\") pod \"22434baf-68e0-4783-8835-27fa1e05b00d\" (UID: \"22434baf-68e0-4783-8835-27fa1e05b00d\") " Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.342128 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22434baf-68e0-4783-8835-27fa1e05b00d-config-volume" (OuterVolumeSpecName: "config-volume") pod "22434baf-68e0-4783-8835-27fa1e05b00d" (UID: "22434baf-68e0-4783-8835-27fa1e05b00d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.353384 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22434baf-68e0-4783-8835-27fa1e05b00d-kube-api-access-vxn7p" (OuterVolumeSpecName: "kube-api-access-vxn7p") pod "22434baf-68e0-4783-8835-27fa1e05b00d" (UID: "22434baf-68e0-4783-8835-27fa1e05b00d"). InnerVolumeSpecName "kube-api-access-vxn7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.369530 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22434baf-68e0-4783-8835-27fa1e05b00d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "22434baf-68e0-4783-8835-27fa1e05b00d" (UID: "22434baf-68e0-4783-8835-27fa1e05b00d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.443307 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxn7p\" (UniqueName: \"kubernetes.io/projected/22434baf-68e0-4783-8835-27fa1e05b00d-kube-api-access-vxn7p\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.443335 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22434baf-68e0-4783-8835-27fa1e05b00d-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.443344 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22434baf-68e0-4783-8835-27fa1e05b00d-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.792753 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.792756 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595" event={"ID":"22434baf-68e0-4783-8835-27fa1e05b00d","Type":"ContainerDied","Data":"818a96eaffb30c26e30541e62b1c2ba1eeff9ae84cc7c97a1e7f2f25dff6485d"} Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.793101 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="818a96eaffb30c26e30541e62b1c2ba1eeff9ae84cc7c97a1e7f2f25dff6485d" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.795141 4741 generic.go:334] "Generic (PLEG): container finished" podID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerID="d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff" exitCode=143 Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.795509 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f1e31bb-969a-4f4b-9826-9b77df8c244b","Type":"ContainerDied","Data":"d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff"} Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.795550 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.795820 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-central-agent" containerID="cri-o://a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" gracePeriod=30 Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.796453 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="sg-core" containerID="cri-o://a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" gracePeriod=30 Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.796465 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="proxy-httpd" containerID="cri-o://ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" gracePeriod=30 Sep 29 19:30:03 crc kubenswrapper[4741]: I0929 19:30:03.796545 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-notification-agent" containerID="cri-o://9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" gracePeriod=30 Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.590975 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.768979 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-scripts\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769357 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-log-httpd\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769399 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-run-httpd\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769497 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-sg-core-conf-yaml\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769538 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2llvp\" (UniqueName: \"kubernetes.io/projected/84280a33-f064-48dc-a074-6283a759cacf-kube-api-access-2llvp\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769580 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-config-data\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769645 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-combined-ca-bundle\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769669 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-ceilometer-tls-certs\") pod \"84280a33-f064-48dc-a074-6283a759cacf\" (UID: \"84280a33-f064-48dc-a074-6283a759cacf\") " Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.769803 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.770034 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.770442 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.775322 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-scripts" (OuterVolumeSpecName: "scripts") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.776070 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84280a33-f064-48dc-a074-6283a759cacf-kube-api-access-2llvp" (OuterVolumeSpecName: "kube-api-access-2llvp") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "kube-api-access-2llvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.797980 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811594 4741 generic.go:334] "Generic (PLEG): container finished" podID="84280a33-f064-48dc-a074-6283a759cacf" containerID="ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" exitCode=0 Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811633 4741 generic.go:334] "Generic (PLEG): container finished" podID="84280a33-f064-48dc-a074-6283a759cacf" containerID="a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" exitCode=2 Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811642 4741 generic.go:334] "Generic (PLEG): container finished" podID="84280a33-f064-48dc-a074-6283a759cacf" containerID="9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" exitCode=0 Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811651 4741 generic.go:334] "Generic (PLEG): container finished" podID="84280a33-f064-48dc-a074-6283a759cacf" containerID="a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" exitCode=0 Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811774 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerDied","Data":"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94"} Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811810 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerDied","Data":"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26"} Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811804 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811829 4741 scope.go:117] "RemoveContainer" containerID="ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811820 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerDied","Data":"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285"} Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811869 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerDied","Data":"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf"} Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.811981 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84280a33-f064-48dc-a074-6283a759cacf","Type":"ContainerDied","Data":"7d0f03bd6f71033fae46dc988d44acc17ad8cc309e903194d3a97de50dd9ccee"} Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.830451 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.839410 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.841771 4741 scope.go:117] "RemoveContainer" containerID="a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.861203 4741 scope.go:117] "RemoveContainer" containerID="9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.872566 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84280a33-f064-48dc-a074-6283a759cacf-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.872593 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.872605 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2llvp\" (UniqueName: \"kubernetes.io/projected/84280a33-f064-48dc-a074-6283a759cacf-kube-api-access-2llvp\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.872614 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.872622 4741 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.872630 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.878533 4741 scope.go:117] "RemoveContainer" containerID="a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.881407 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-config-data" (OuterVolumeSpecName: "config-data") pod "84280a33-f064-48dc-a074-6283a759cacf" (UID: "84280a33-f064-48dc-a074-6283a759cacf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.897192 4741 scope.go:117] "RemoveContainer" containerID="ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" Sep 29 19:30:04 crc kubenswrapper[4741]: E0929 19:30:04.897570 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": container with ID starting with ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94 not found: ID does not exist" containerID="ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.897603 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94"} err="failed to get container status \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": rpc error: code = NotFound desc = could not find container \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": container with ID starting with ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.897624 4741 scope.go:117] "RemoveContainer" containerID="a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" Sep 29 19:30:04 crc kubenswrapper[4741]: E0929 19:30:04.898158 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": container with ID starting with a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26 not found: ID does not exist" containerID="a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.898184 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26"} err="failed to get container status \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": rpc error: code = NotFound desc = could not find container \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": container with ID starting with a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.898197 4741 scope.go:117] "RemoveContainer" containerID="9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" Sep 29 19:30:04 crc kubenswrapper[4741]: E0929 19:30:04.898420 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": container with ID starting with 9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285 not found: ID does not exist" containerID="9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.898445 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285"} err="failed to get container status \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": rpc error: code = NotFound desc = could not find container \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": container with ID starting with 9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.898477 4741 scope.go:117] "RemoveContainer" containerID="a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" Sep 29 19:30:04 crc kubenswrapper[4741]: E0929 19:30:04.898843 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": container with ID starting with a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf not found: ID does not exist" containerID="a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.898866 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf"} err="failed to get container status \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": rpc error: code = NotFound desc = could not find container \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": container with ID starting with a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.898879 4741 scope.go:117] "RemoveContainer" containerID="ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.903144 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94"} err="failed to get container status \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": rpc error: code = NotFound desc = could not find container \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": container with ID starting with ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.903200 4741 scope.go:117] "RemoveContainer" containerID="a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.903648 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26"} err="failed to get container status \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": rpc error: code = NotFound desc = could not find container \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": container with ID starting with a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.903683 4741 scope.go:117] "RemoveContainer" containerID="9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.903933 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285"} err="failed to get container status \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": rpc error: code = NotFound desc = could not find container \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": container with ID starting with 9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.903959 4741 scope.go:117] "RemoveContainer" containerID="a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.904143 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf"} err="failed to get container status \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": rpc error: code = NotFound desc = could not find container \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": container with ID starting with a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.904170 4741 scope.go:117] "RemoveContainer" containerID="ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.904587 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94"} err="failed to get container status \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": rpc error: code = NotFound desc = could not find container \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": container with ID starting with ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.904614 4741 scope.go:117] "RemoveContainer" containerID="a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.904794 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26"} err="failed to get container status \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": rpc error: code = NotFound desc = could not find container \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": container with ID starting with a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.904813 4741 scope.go:117] "RemoveContainer" containerID="9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905116 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285"} err="failed to get container status \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": rpc error: code = NotFound desc = could not find container \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": container with ID starting with 9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905146 4741 scope.go:117] "RemoveContainer" containerID="a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905321 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf"} err="failed to get container status \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": rpc error: code = NotFound desc = could not find container \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": container with ID starting with a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905337 4741 scope.go:117] "RemoveContainer" containerID="ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905506 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94"} err="failed to get container status \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": rpc error: code = NotFound desc = could not find container \"ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94\": container with ID starting with ba5dd71fc37808052e2006789a2e09ab78a478569377b600df23a9d6ab200c94 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905520 4741 scope.go:117] "RemoveContainer" containerID="a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905767 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26"} err="failed to get container status \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": rpc error: code = NotFound desc = could not find container \"a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26\": container with ID starting with a44beb7ba1e5c888f02a7016d3e4c0f389a54eb14fe387cea57ed5294ebffb26 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.905797 4741 scope.go:117] "RemoveContainer" containerID="9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.906287 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285"} err="failed to get container status \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": rpc error: code = NotFound desc = could not find container \"9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285\": container with ID starting with 9049e1d5fd517667b03337665e732a4dc45bc76fab7332f98e276de8409aa285 not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.906371 4741 scope.go:117] "RemoveContainer" containerID="a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.906897 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf"} err="failed to get container status \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": rpc error: code = NotFound desc = could not find container \"a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf\": container with ID starting with a01f15644cc5655192695cc1fc2526ecac5d36782bc49e559c44ae9bbf3f96bf not found: ID does not exist" Sep 29 19:30:04 crc kubenswrapper[4741]: I0929 19:30:04.973858 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84280a33-f064-48dc-a074-6283a759cacf-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.137826 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.146797 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.158483 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:30:05 crc kubenswrapper[4741]: E0929 19:30:05.159074 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22434baf-68e0-4783-8835-27fa1e05b00d" containerName="collect-profiles" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159096 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="22434baf-68e0-4783-8835-27fa1e05b00d" containerName="collect-profiles" Sep 29 19:30:05 crc kubenswrapper[4741]: E0929 19:30:05.159126 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-notification-agent" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159137 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-notification-agent" Sep 29 19:30:05 crc kubenswrapper[4741]: E0929 19:30:05.159147 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-central-agent" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159152 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-central-agent" Sep 29 19:30:05 crc kubenswrapper[4741]: E0929 19:30:05.159164 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="sg-core" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159170 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="sg-core" Sep 29 19:30:05 crc kubenswrapper[4741]: E0929 19:30:05.159189 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="proxy-httpd" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159195 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="proxy-httpd" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159369 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-notification-agent" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159380 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="proxy-httpd" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159919 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="sg-core" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159949 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="22434baf-68e0-4783-8835-27fa1e05b00d" containerName="collect-profiles" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.159962 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="84280a33-f064-48dc-a074-6283a759cacf" containerName="ceilometer-central-agent" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.161765 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.164036 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.164218 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.174970 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.175138 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285018 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhs7q\" (UniqueName: \"kubernetes.io/projected/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-kube-api-access-hhs7q\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285088 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-run-httpd\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285167 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-scripts\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285184 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285249 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-config-data\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285278 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-log-httpd\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285344 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.285410 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386736 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386794 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386835 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhs7q\" (UniqueName: \"kubernetes.io/projected/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-kube-api-access-hhs7q\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386864 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-run-httpd\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386897 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-scripts\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386916 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386950 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-config-data\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.386981 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-log-httpd\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.387411 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-run-httpd\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.387423 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-log-httpd\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.390180 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.390570 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.390715 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-config-data\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.391104 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-scripts\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.391601 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.405047 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhs7q\" (UniqueName: \"kubernetes.io/projected/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-kube-api-access-hhs7q\") pod \"ceilometer-0\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.498757 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:30:05 crc kubenswrapper[4741]: I0929 19:30:05.906217 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:30:05 crc kubenswrapper[4741]: W0929 19:30:05.913030 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod875d9bb5_32f8_44c7_aa70_0dfcaf8f9834.slice/crio-9ecb21a53a3b4d17f2a69f3f44646b8f16e0c8ee781bf24d844659ca14f7294a WatchSource:0}: Error finding container 9ecb21a53a3b4d17f2a69f3f44646b8f16e0c8ee781bf24d844659ca14f7294a: Status 404 returned error can't find the container with id 9ecb21a53a3b4d17f2a69f3f44646b8f16e0c8ee781bf24d844659ca14f7294a Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.365530 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.509277 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-config-data\") pod \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.509597 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-combined-ca-bundle\") pod \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.509730 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1e31bb-969a-4f4b-9826-9b77df8c244b-logs\") pod \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.509826 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsc7n\" (UniqueName: \"kubernetes.io/projected/2f1e31bb-969a-4f4b-9826-9b77df8c244b-kube-api-access-jsc7n\") pod \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\" (UID: \"2f1e31bb-969a-4f4b-9826-9b77df8c244b\") " Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.510375 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f1e31bb-969a-4f4b-9826-9b77df8c244b-logs" (OuterVolumeSpecName: "logs") pod "2f1e31bb-969a-4f4b-9826-9b77df8c244b" (UID: "2f1e31bb-969a-4f4b-9826-9b77df8c244b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.518596 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f1e31bb-969a-4f4b-9826-9b77df8c244b-kube-api-access-jsc7n" (OuterVolumeSpecName: "kube-api-access-jsc7n") pod "2f1e31bb-969a-4f4b-9826-9b77df8c244b" (UID: "2f1e31bb-969a-4f4b-9826-9b77df8c244b"). InnerVolumeSpecName "kube-api-access-jsc7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.550784 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f1e31bb-969a-4f4b-9826-9b77df8c244b" (UID: "2f1e31bb-969a-4f4b-9826-9b77df8c244b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.556467 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-config-data" (OuterVolumeSpecName: "config-data") pod "2f1e31bb-969a-4f4b-9826-9b77df8c244b" (UID: "2f1e31bb-969a-4f4b-9826-9b77df8c244b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.612252 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1e31bb-969a-4f4b-9826-9b77df8c244b-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.612295 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsc7n\" (UniqueName: \"kubernetes.io/projected/2f1e31bb-969a-4f4b-9826-9b77df8c244b-kube-api-access-jsc7n\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.612305 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.612313 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1e31bb-969a-4f4b-9826-9b77df8c244b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.844332 4741 generic.go:334] "Generic (PLEG): container finished" podID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerID="0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf" exitCode=0 Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.844385 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.844510 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f1e31bb-969a-4f4b-9826-9b77df8c244b","Type":"ContainerDied","Data":"0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf"} Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.844557 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f1e31bb-969a-4f4b-9826-9b77df8c244b","Type":"ContainerDied","Data":"d9e4660888a9955b1cbbd75b284d20d2ca3a721155b7508dde852eef26bec17a"} Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.844585 4741 scope.go:117] "RemoveContainer" containerID="0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.845890 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerStarted","Data":"9ecb21a53a3b4d17f2a69f3f44646b8f16e0c8ee781bf24d844659ca14f7294a"} Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.866807 4741 scope.go:117] "RemoveContainer" containerID="d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.895393 4741 scope.go:117] "RemoveContainer" containerID="0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.896060 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:06 crc kubenswrapper[4741]: E0929 19:30:06.896147 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf\": container with ID starting with 0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf not found: ID does not exist" containerID="0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.896181 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf"} err="failed to get container status \"0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf\": rpc error: code = NotFound desc = could not find container \"0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf\": container with ID starting with 0c6a620639d392059724fe269e43fa2d2acbb35f5fb10290018b18bfb5c6abdf not found: ID does not exist" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.896201 4741 scope.go:117] "RemoveContainer" containerID="d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff" Sep 29 19:30:06 crc kubenswrapper[4741]: E0929 19:30:06.897644 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff\": container with ID starting with d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff not found: ID does not exist" containerID="d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.897693 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff"} err="failed to get container status \"d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff\": rpc error: code = NotFound desc = could not find container \"d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff\": container with ID starting with d809f418d93eeb1a9dc3afae338170acad13205dcdfc6814a3f4fa4a41a5c7ff not found: ID does not exist" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.915839 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.923842 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:06 crc kubenswrapper[4741]: E0929 19:30:06.924375 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-api" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.924401 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-api" Sep 29 19:30:06 crc kubenswrapper[4741]: E0929 19:30:06.924460 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-log" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.924469 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-log" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.924750 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-log" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.924778 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" containerName="nova-api-api" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.926051 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.928109 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.928298 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.929492 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 29 19:30:06 crc kubenswrapper[4741]: I0929 19:30:06.931267 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.020592 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs9gl\" (UniqueName: \"kubernetes.io/projected/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-kube-api-access-qs9gl\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.020676 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.020716 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-logs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.020783 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.020938 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-config-data\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.021180 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-public-tls-certs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.099710 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f1e31bb-969a-4f4b-9826-9b77df8c244b" path="/var/lib/kubelet/pods/2f1e31bb-969a-4f4b-9826-9b77df8c244b/volumes" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.100387 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84280a33-f064-48dc-a074-6283a759cacf" path="/var/lib/kubelet/pods/84280a33-f064-48dc-a074-6283a759cacf/volumes" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.115482 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.122806 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs9gl\" (UniqueName: \"kubernetes.io/projected/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-kube-api-access-qs9gl\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.122843 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.122877 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-logs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.122928 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.122955 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-config-data\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.123002 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-public-tls-certs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.123931 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-logs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.128349 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-config-data\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.128413 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-public-tls-certs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.130009 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.135459 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.136798 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.141088 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs9gl\" (UniqueName: \"kubernetes.io/projected/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-kube-api-access-qs9gl\") pod \"nova-api-0\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.306150 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.739790 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:07 crc kubenswrapper[4741]: W0929 19:30:07.742565 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60bef6d7_0583_4f73_97c6_b9b5dafb45d8.slice/crio-6da1f419fa1457eb9ee665dfb7f307e6e300248ba7861dcb67d53a51415459a7 WatchSource:0}: Error finding container 6da1f419fa1457eb9ee665dfb7f307e6e300248ba7861dcb67d53a51415459a7: Status 404 returned error can't find the container with id 6da1f419fa1457eb9ee665dfb7f307e6e300248ba7861dcb67d53a51415459a7 Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.855336 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bef6d7-0583-4f73-97c6-b9b5dafb45d8","Type":"ContainerStarted","Data":"6da1f419fa1457eb9ee665dfb7f307e6e300248ba7861dcb67d53a51415459a7"} Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.856987 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerStarted","Data":"3d41a9cc13ba07eab14e449222ca1eba4bc2f913c9d707b7f2f9fe2953779503"} Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.857009 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerStarted","Data":"64aaaacdbcb9f52035038a67edb1624996ab0fb313a96a375112066aa4d0c3a6"} Sep 29 19:30:07 crc kubenswrapper[4741]: I0929 19:30:07.876058 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.069630 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dbj47"] Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.072279 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.075120 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.075379 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.076187 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dbj47"] Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.143777 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.143874 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-config-data\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.143963 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkxv9\" (UniqueName: \"kubernetes.io/projected/acf52497-97c7-4e1f-a79c-29ec872359c5-kube-api-access-vkxv9\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.143993 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-scripts\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.245580 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkxv9\" (UniqueName: \"kubernetes.io/projected/acf52497-97c7-4e1f-a79c-29ec872359c5-kube-api-access-vkxv9\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.245945 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-scripts\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.246040 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.246125 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-config-data\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.250336 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-config-data\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.254879 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.255793 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-scripts\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.269278 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkxv9\" (UniqueName: \"kubernetes.io/projected/acf52497-97c7-4e1f-a79c-29ec872359c5-kube-api-access-vkxv9\") pod \"nova-cell1-cell-mapping-dbj47\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.418738 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:08 crc kubenswrapper[4741]: W0929 19:30:08.871644 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacf52497_97c7_4e1f_a79c_29ec872359c5.slice/crio-b09433b7242e7e05c364c4c4474237b01850dbe3ddf81d59f622bd945f2ce290 WatchSource:0}: Error finding container b09433b7242e7e05c364c4c4474237b01850dbe3ddf81d59f622bd945f2ce290: Status 404 returned error can't find the container with id b09433b7242e7e05c364c4c4474237b01850dbe3ddf81d59f622bd945f2ce290 Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.873983 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bef6d7-0583-4f73-97c6-b9b5dafb45d8","Type":"ContainerStarted","Data":"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f"} Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.874020 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bef6d7-0583-4f73-97c6-b9b5dafb45d8","Type":"ContainerStarted","Data":"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967"} Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.875026 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dbj47"] Sep 29 19:30:08 crc kubenswrapper[4741]: I0929 19:30:08.876818 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerStarted","Data":"113041bee656cb97e263bb0fbf9286ccb77bdf43fa28dcf029dc790fd12a60d8"} Sep 29 19:30:09 crc kubenswrapper[4741]: I0929 19:30:09.138994 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.138975946 podStartE2EDuration="3.138975946s" podCreationTimestamp="2025-09-29 19:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:30:08.897264646 +0000 UTC m=+1250.545053998" watchObservedRunningTime="2025-09-29 19:30:09.138975946 +0000 UTC m=+1250.786765278" Sep 29 19:30:09 crc kubenswrapper[4741]: I0929 19:30:09.891901 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dbj47" event={"ID":"acf52497-97c7-4e1f-a79c-29ec872359c5","Type":"ContainerStarted","Data":"07ff09e353f0b9f0d4e0a433a86e107581fb646712047e332b67415755e5ec2d"} Sep 29 19:30:09 crc kubenswrapper[4741]: I0929 19:30:09.892150 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dbj47" event={"ID":"acf52497-97c7-4e1f-a79c-29ec872359c5","Type":"ContainerStarted","Data":"b09433b7242e7e05c364c4c4474237b01850dbe3ddf81d59f622bd945f2ce290"} Sep 29 19:30:09 crc kubenswrapper[4741]: I0929 19:30:09.912162 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dbj47" podStartSLOduration=1.912140537 podStartE2EDuration="1.912140537s" podCreationTimestamp="2025-09-29 19:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:30:09.903999041 +0000 UTC m=+1251.551788423" watchObservedRunningTime="2025-09-29 19:30:09.912140537 +0000 UTC m=+1251.559929869" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.237613 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.307505 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9nm5f"] Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.307737 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" podUID="f0900a38-6648-4c90-9458-c4395b446a3a" containerName="dnsmasq-dns" containerID="cri-o://816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11" gracePeriod=10 Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.845670 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.907559 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerStarted","Data":"167d2c7f919eb9ab445a2017a6763de4f7c20dd26fb277979f6106c71a0731b5"} Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.909640 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.915259 4741 generic.go:334] "Generic (PLEG): container finished" podID="f0900a38-6648-4c90-9458-c4395b446a3a" containerID="816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11" exitCode=0 Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.916476 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.916748 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" event={"ID":"f0900a38-6648-4c90-9458-c4395b446a3a","Type":"ContainerDied","Data":"816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11"} Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.916836 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9nm5f" event={"ID":"f0900a38-6648-4c90-9458-c4395b446a3a","Type":"ContainerDied","Data":"af8964bb39257bb53d73fa5b041b8f800b07703ce84bcfbefc2583cce8891e0d"} Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.916859 4741 scope.go:117] "RemoveContainer" containerID="816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.944219 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.100916053 podStartE2EDuration="5.944199279s" podCreationTimestamp="2025-09-29 19:30:05 +0000 UTC" firstStartedPulling="2025-09-29 19:30:05.915695115 +0000 UTC m=+1247.563484447" lastFinishedPulling="2025-09-29 19:30:09.758978341 +0000 UTC m=+1251.406767673" observedRunningTime="2025-09-29 19:30:10.933462911 +0000 UTC m=+1252.581252273" watchObservedRunningTime="2025-09-29 19:30:10.944199279 +0000 UTC m=+1252.591988621" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.961967 4741 scope.go:117] "RemoveContainer" containerID="d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.998484 4741 scope.go:117] "RemoveContainer" containerID="816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11" Sep 29 19:30:10 crc kubenswrapper[4741]: E0929 19:30:10.998825 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11\": container with ID starting with 816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11 not found: ID does not exist" containerID="816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.998864 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11"} err="failed to get container status \"816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11\": rpc error: code = NotFound desc = could not find container \"816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11\": container with ID starting with 816d57bea48d993cbc1edc0809d9f5a855a51b274fbbd94f2ac9d4ccad166c11 not found: ID does not exist" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.998886 4741 scope.go:117] "RemoveContainer" containerID="d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf" Sep 29 19:30:10 crc kubenswrapper[4741]: E0929 19:30:10.999084 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf\": container with ID starting with d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf not found: ID does not exist" containerID="d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf" Sep 29 19:30:10 crc kubenswrapper[4741]: I0929 19:30:10.999114 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf"} err="failed to get container status \"d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf\": rpc error: code = NotFound desc = could not find container \"d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf\": container with ID starting with d317619ebdcff12f3fcedd44b2dc608f1a94c3114e4c6dcf1b48279ac77403cf not found: ID does not exist" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.004081 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-config\") pod \"f0900a38-6648-4c90-9458-c4395b446a3a\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.004193 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-swift-storage-0\") pod \"f0900a38-6648-4c90-9458-c4395b446a3a\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.004349 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-sb\") pod \"f0900a38-6648-4c90-9458-c4395b446a3a\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.004949 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-nb\") pod \"f0900a38-6648-4c90-9458-c4395b446a3a\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.005032 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m474\" (UniqueName: \"kubernetes.io/projected/f0900a38-6648-4c90-9458-c4395b446a3a-kube-api-access-2m474\") pod \"f0900a38-6648-4c90-9458-c4395b446a3a\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.005318 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-svc\") pod \"f0900a38-6648-4c90-9458-c4395b446a3a\" (UID: \"f0900a38-6648-4c90-9458-c4395b446a3a\") " Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.015298 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0900a38-6648-4c90-9458-c4395b446a3a-kube-api-access-2m474" (OuterVolumeSpecName: "kube-api-access-2m474") pod "f0900a38-6648-4c90-9458-c4395b446a3a" (UID: "f0900a38-6648-4c90-9458-c4395b446a3a"). InnerVolumeSpecName "kube-api-access-2m474". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.065233 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-config" (OuterVolumeSpecName: "config") pod "f0900a38-6648-4c90-9458-c4395b446a3a" (UID: "f0900a38-6648-4c90-9458-c4395b446a3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.067836 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f0900a38-6648-4c90-9458-c4395b446a3a" (UID: "f0900a38-6648-4c90-9458-c4395b446a3a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.068475 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f0900a38-6648-4c90-9458-c4395b446a3a" (UID: "f0900a38-6648-4c90-9458-c4395b446a3a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.081485 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f0900a38-6648-4c90-9458-c4395b446a3a" (UID: "f0900a38-6648-4c90-9458-c4395b446a3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.090645 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f0900a38-6648-4c90-9458-c4395b446a3a" (UID: "f0900a38-6648-4c90-9458-c4395b446a3a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.110780 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.110825 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.110841 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.110851 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.110865 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m474\" (UniqueName: \"kubernetes.io/projected/f0900a38-6648-4c90-9458-c4395b446a3a-kube-api-access-2m474\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.110873 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0900a38-6648-4c90-9458-c4395b446a3a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.245709 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9nm5f"] Sep 29 19:30:11 crc kubenswrapper[4741]: I0929 19:30:11.254652 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9nm5f"] Sep 29 19:30:13 crc kubenswrapper[4741]: I0929 19:30:13.100922 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0900a38-6648-4c90-9458-c4395b446a3a" path="/var/lib/kubelet/pods/f0900a38-6648-4c90-9458-c4395b446a3a/volumes" Sep 29 19:30:13 crc kubenswrapper[4741]: I0929 19:30:13.945354 4741 generic.go:334] "Generic (PLEG): container finished" podID="acf52497-97c7-4e1f-a79c-29ec872359c5" containerID="07ff09e353f0b9f0d4e0a433a86e107581fb646712047e332b67415755e5ec2d" exitCode=0 Sep 29 19:30:13 crc kubenswrapper[4741]: I0929 19:30:13.945510 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dbj47" event={"ID":"acf52497-97c7-4e1f-a79c-29ec872359c5","Type":"ContainerDied","Data":"07ff09e353f0b9f0d4e0a433a86e107581fb646712047e332b67415755e5ec2d"} Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.364231 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.509742 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-scripts\") pod \"acf52497-97c7-4e1f-a79c-29ec872359c5\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.510177 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-config-data\") pod \"acf52497-97c7-4e1f-a79c-29ec872359c5\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.510331 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkxv9\" (UniqueName: \"kubernetes.io/projected/acf52497-97c7-4e1f-a79c-29ec872359c5-kube-api-access-vkxv9\") pod \"acf52497-97c7-4e1f-a79c-29ec872359c5\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.510440 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-combined-ca-bundle\") pod \"acf52497-97c7-4e1f-a79c-29ec872359c5\" (UID: \"acf52497-97c7-4e1f-a79c-29ec872359c5\") " Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.515933 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acf52497-97c7-4e1f-a79c-29ec872359c5-kube-api-access-vkxv9" (OuterVolumeSpecName: "kube-api-access-vkxv9") pod "acf52497-97c7-4e1f-a79c-29ec872359c5" (UID: "acf52497-97c7-4e1f-a79c-29ec872359c5"). InnerVolumeSpecName "kube-api-access-vkxv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.516718 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-scripts" (OuterVolumeSpecName: "scripts") pod "acf52497-97c7-4e1f-a79c-29ec872359c5" (UID: "acf52497-97c7-4e1f-a79c-29ec872359c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.541088 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acf52497-97c7-4e1f-a79c-29ec872359c5" (UID: "acf52497-97c7-4e1f-a79c-29ec872359c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.544682 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-config-data" (OuterVolumeSpecName: "config-data") pod "acf52497-97c7-4e1f-a79c-29ec872359c5" (UID: "acf52497-97c7-4e1f-a79c-29ec872359c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.612756 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.612808 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkxv9\" (UniqueName: \"kubernetes.io/projected/acf52497-97c7-4e1f-a79c-29ec872359c5-kube-api-access-vkxv9\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.612828 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.612847 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acf52497-97c7-4e1f-a79c-29ec872359c5-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.968787 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dbj47" event={"ID":"acf52497-97c7-4e1f-a79c-29ec872359c5","Type":"ContainerDied","Data":"b09433b7242e7e05c364c4c4474237b01850dbe3ddf81d59f622bd945f2ce290"} Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.968832 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b09433b7242e7e05c364c4c4474237b01850dbe3ddf81d59f622bd945f2ce290" Sep 29 19:30:15 crc kubenswrapper[4741]: I0929 19:30:15.968930 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dbj47" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.229212 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.229805 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-log" containerID="cri-o://bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967" gracePeriod=30 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.230369 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-api" containerID="cri-o://0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f" gracePeriod=30 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.248797 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.249477 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-log" containerID="cri-o://8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1" gracePeriod=30 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.250126 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-metadata" containerID="cri-o://730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6" gracePeriod=30 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.263358 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.263648 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cc0d61fc-e880-4857-bdea-b90f2a86b4c2" containerName="nova-scheduler-scheduler" containerID="cri-o://7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4" gracePeriod=30 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.824689 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.840266 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-internal-tls-certs\") pod \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.840361 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs9gl\" (UniqueName: \"kubernetes.io/projected/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-kube-api-access-qs9gl\") pod \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.840463 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-public-tls-certs\") pod \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.840496 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-logs\") pod \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.840597 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-config-data\") pod \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.840666 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-combined-ca-bundle\") pod \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\" (UID: \"60bef6d7-0583-4f73-97c6-b9b5dafb45d8\") " Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.845004 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-logs" (OuterVolumeSpecName: "logs") pod "60bef6d7-0583-4f73-97c6-b9b5dafb45d8" (UID: "60bef6d7-0583-4f73-97c6-b9b5dafb45d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.848342 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-kube-api-access-qs9gl" (OuterVolumeSpecName: "kube-api-access-qs9gl") pod "60bef6d7-0583-4f73-97c6-b9b5dafb45d8" (UID: "60bef6d7-0583-4f73-97c6-b9b5dafb45d8"). InnerVolumeSpecName "kube-api-access-qs9gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.875971 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-config-data" (OuterVolumeSpecName: "config-data") pod "60bef6d7-0583-4f73-97c6-b9b5dafb45d8" (UID: "60bef6d7-0583-4f73-97c6-b9b5dafb45d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.883836 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60bef6d7-0583-4f73-97c6-b9b5dafb45d8" (UID: "60bef6d7-0583-4f73-97c6-b9b5dafb45d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.897736 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "60bef6d7-0583-4f73-97c6-b9b5dafb45d8" (UID: "60bef6d7-0583-4f73-97c6-b9b5dafb45d8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.939596 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "60bef6d7-0583-4f73-97c6-b9b5dafb45d8" (UID: "60bef6d7-0583-4f73-97c6-b9b5dafb45d8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.943969 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.944003 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.944023 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.944041 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs9gl\" (UniqueName: \"kubernetes.io/projected/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-kube-api-access-qs9gl\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.944059 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.944073 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bef6d7-0583-4f73-97c6-b9b5dafb45d8-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.984516 4741 generic.go:334] "Generic (PLEG): container finished" podID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerID="8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1" exitCode=143 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.984602 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4","Type":"ContainerDied","Data":"8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1"} Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.988039 4741 generic.go:334] "Generic (PLEG): container finished" podID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerID="0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f" exitCode=0 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.988073 4741 generic.go:334] "Generic (PLEG): container finished" podID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerID="bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967" exitCode=143 Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.988095 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.988098 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bef6d7-0583-4f73-97c6-b9b5dafb45d8","Type":"ContainerDied","Data":"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f"} Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.988197 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bef6d7-0583-4f73-97c6-b9b5dafb45d8","Type":"ContainerDied","Data":"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967"} Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.988212 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bef6d7-0583-4f73-97c6-b9b5dafb45d8","Type":"ContainerDied","Data":"6da1f419fa1457eb9ee665dfb7f307e6e300248ba7861dcb67d53a51415459a7"} Sep 29 19:30:16 crc kubenswrapper[4741]: I0929 19:30:16.988231 4741 scope.go:117] "RemoveContainer" containerID="0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.022236 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.024271 4741 scope.go:117] "RemoveContainer" containerID="bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.038226 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.045161 4741 scope.go:117] "RemoveContainer" containerID="0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f" Sep 29 19:30:17 crc kubenswrapper[4741]: E0929 19:30:17.046091 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f\": container with ID starting with 0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f not found: ID does not exist" containerID="0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.046141 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f"} err="failed to get container status \"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f\": rpc error: code = NotFound desc = could not find container \"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f\": container with ID starting with 0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f not found: ID does not exist" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.046175 4741 scope.go:117] "RemoveContainer" containerID="bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967" Sep 29 19:30:17 crc kubenswrapper[4741]: E0929 19:30:17.048677 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967\": container with ID starting with bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967 not found: ID does not exist" containerID="bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.048715 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967"} err="failed to get container status \"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967\": rpc error: code = NotFound desc = could not find container \"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967\": container with ID starting with bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967 not found: ID does not exist" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.048737 4741 scope.go:117] "RemoveContainer" containerID="0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.049108 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f"} err="failed to get container status \"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f\": rpc error: code = NotFound desc = could not find container \"0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f\": container with ID starting with 0d82d9f461ec5a2aa287927d1eddc9380380dadd1a59e0a5b41f2bfdd292231f not found: ID does not exist" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.049150 4741 scope.go:117] "RemoveContainer" containerID="bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.049509 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967"} err="failed to get container status \"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967\": rpc error: code = NotFound desc = could not find container \"bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967\": container with ID starting with bf4362bd38dfbf935cd4deca2412069f9804ba2e1456c9a492944e9d4dc59967 not found: ID does not exist" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056060 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:17 crc kubenswrapper[4741]: E0929 19:30:17.056371 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-api" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056392 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-api" Sep 29 19:30:17 crc kubenswrapper[4741]: E0929 19:30:17.056421 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0900a38-6648-4c90-9458-c4395b446a3a" containerName="dnsmasq-dns" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056429 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0900a38-6648-4c90-9458-c4395b446a3a" containerName="dnsmasq-dns" Sep 29 19:30:17 crc kubenswrapper[4741]: E0929 19:30:17.056446 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-log" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056451 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-log" Sep 29 19:30:17 crc kubenswrapper[4741]: E0929 19:30:17.056472 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0900a38-6648-4c90-9458-c4395b446a3a" containerName="init" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056480 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0900a38-6648-4c90-9458-c4395b446a3a" containerName="init" Sep 29 19:30:17 crc kubenswrapper[4741]: E0929 19:30:17.056496 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf52497-97c7-4e1f-a79c-29ec872359c5" containerName="nova-manage" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056503 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf52497-97c7-4e1f-a79c-29ec872359c5" containerName="nova-manage" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056694 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-api" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056705 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" containerName="nova-api-log" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056717 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0900a38-6648-4c90-9458-c4395b446a3a" containerName="dnsmasq-dns" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.056727 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="acf52497-97c7-4e1f-a79c-29ec872359c5" containerName="nova-manage" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.058083 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.060187 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.061231 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.061357 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.066707 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.104757 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60bef6d7-0583-4f73-97c6-b9b5dafb45d8" path="/var/lib/kubelet/pods/60bef6d7-0583-4f73-97c6-b9b5dafb45d8/volumes" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.147054 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-public-tls-certs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.147109 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.147138 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljjsz\" (UniqueName: \"kubernetes.io/projected/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-kube-api-access-ljjsz\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.147183 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-config-data\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.147205 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.147365 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-logs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.248110 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljjsz\" (UniqueName: \"kubernetes.io/projected/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-kube-api-access-ljjsz\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.248177 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-config-data\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.248205 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.248281 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-logs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.248304 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-public-tls-certs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.248340 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.248715 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-logs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.251706 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.252069 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-public-tls-certs\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.254930 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.255043 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-config-data\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.261422 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljjsz\" (UniqueName: \"kubernetes.io/projected/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-kube-api-access-ljjsz\") pod \"nova-api-0\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.407735 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.682851 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.756618 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlvj5\" (UniqueName: \"kubernetes.io/projected/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-kube-api-access-nlvj5\") pod \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.757125 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-combined-ca-bundle\") pod \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.757159 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-config-data\") pod \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\" (UID: \"cc0d61fc-e880-4857-bdea-b90f2a86b4c2\") " Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.762294 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-kube-api-access-nlvj5" (OuterVolumeSpecName: "kube-api-access-nlvj5") pod "cc0d61fc-e880-4857-bdea-b90f2a86b4c2" (UID: "cc0d61fc-e880-4857-bdea-b90f2a86b4c2"). InnerVolumeSpecName "kube-api-access-nlvj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.787193 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc0d61fc-e880-4857-bdea-b90f2a86b4c2" (UID: "cc0d61fc-e880-4857-bdea-b90f2a86b4c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.800631 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-config-data" (OuterVolumeSpecName: "config-data") pod "cc0d61fc-e880-4857-bdea-b90f2a86b4c2" (UID: "cc0d61fc-e880-4857-bdea-b90f2a86b4c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.859473 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.859530 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlvj5\" (UniqueName: \"kubernetes.io/projected/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-kube-api-access-nlvj5\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.859549 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d61fc-e880-4857-bdea-b90f2a86b4c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:17 crc kubenswrapper[4741]: I0929 19:30:17.916116 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:30:17 crc kubenswrapper[4741]: W0929 19:30:17.924146 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe00d3f7_3dee_4e64_8559_2a5efe4ceba3.slice/crio-d0f7b1610351d19418e7412967a6095f0b940278fd9aaf2840335b3e894dc531 WatchSource:0}: Error finding container d0f7b1610351d19418e7412967a6095f0b940278fd9aaf2840335b3e894dc531: Status 404 returned error can't find the container with id d0f7b1610351d19418e7412967a6095f0b940278fd9aaf2840335b3e894dc531 Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.028226 4741 generic.go:334] "Generic (PLEG): container finished" podID="cc0d61fc-e880-4857-bdea-b90f2a86b4c2" containerID="7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4" exitCode=0 Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.028435 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc0d61fc-e880-4857-bdea-b90f2a86b4c2","Type":"ContainerDied","Data":"7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4"} Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.028558 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc0d61fc-e880-4857-bdea-b90f2a86b4c2","Type":"ContainerDied","Data":"9cd20a1e71ce225dec4b19917b45293e7176f419aad8b03b7a49b4339d607779"} Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.028640 4741 scope.go:117] "RemoveContainer" containerID="7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.028821 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.062503 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be00d3f7-3dee-4e64-8559-2a5efe4ceba3","Type":"ContainerStarted","Data":"d0f7b1610351d19418e7412967a6095f0b940278fd9aaf2840335b3e894dc531"} Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.083660 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.095777 4741 scope.go:117] "RemoveContainer" containerID="7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.103861 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:30:18 crc kubenswrapper[4741]: E0929 19:30:18.104884 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4\": container with ID starting with 7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4 not found: ID does not exist" containerID="7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.104934 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4"} err="failed to get container status \"7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4\": rpc error: code = NotFound desc = could not find container \"7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4\": container with ID starting with 7fa2b0fe683fa9cf4e282d3f15040ffc3090f892edb5acd39749c76eaec8e7e4 not found: ID does not exist" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.113868 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:30:18 crc kubenswrapper[4741]: E0929 19:30:18.114277 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0d61fc-e880-4857-bdea-b90f2a86b4c2" containerName="nova-scheduler-scheduler" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.114293 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0d61fc-e880-4857-bdea-b90f2a86b4c2" containerName="nova-scheduler-scheduler" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.114493 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0d61fc-e880-4857-bdea-b90f2a86b4c2" containerName="nova-scheduler-scheduler" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.115037 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.122289 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.132205 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.169221 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-config-data\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.169575 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q8lc\" (UniqueName: \"kubernetes.io/projected/9d18b4c1-d423-4840-97ff-c322272c3aa3-kube-api-access-2q8lc\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.169614 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.270914 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-config-data\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.271007 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q8lc\" (UniqueName: \"kubernetes.io/projected/9d18b4c1-d423-4840-97ff-c322272c3aa3-kube-api-access-2q8lc\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.271041 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.276620 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-config-data\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.276638 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.285271 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q8lc\" (UniqueName: \"kubernetes.io/projected/9d18b4c1-d423-4840-97ff-c322272c3aa3-kube-api-access-2q8lc\") pod \"nova-scheduler-0\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.432864 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:30:18 crc kubenswrapper[4741]: I0929 19:30:18.894494 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.073386 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d18b4c1-d423-4840-97ff-c322272c3aa3","Type":"ContainerStarted","Data":"893ae0f43d22e06b970470419b3519a6562eae482a28106ea379115dd697272d"} Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.076854 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be00d3f7-3dee-4e64-8559-2a5efe4ceba3","Type":"ContainerStarted","Data":"040aa5ec5b498aa5b60ba516e013b9839f5b611e8d7d5311e0be82ba0ff8b3c7"} Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.076928 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be00d3f7-3dee-4e64-8559-2a5efe4ceba3","Type":"ContainerStarted","Data":"8dccd7363f3ef4f5fc4ea08cd195ce440540c0c08f0954b35a076493bfa65bf4"} Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.102782 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.102763573 podStartE2EDuration="2.102763573s" podCreationTimestamp="2025-09-29 19:30:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:30:19.094787153 +0000 UTC m=+1260.742576515" watchObservedRunningTime="2025-09-29 19:30:19.102763573 +0000 UTC m=+1260.750552905" Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.110938 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0d61fc-e880-4857-bdea-b90f2a86b4c2" path="/var/lib/kubelet/pods/cc0d61fc-e880-4857-bdea-b90f2a86b4c2/volumes" Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.410886 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": read tcp 10.217.0.2:57722->10.217.0.187:8775: read: connection reset by peer" Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.410950 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": read tcp 10.217.0.2:57726->10.217.0.187:8775: read: connection reset by peer" Sep 29 19:30:19 crc kubenswrapper[4741]: I0929 19:30:19.901329 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.000229 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-combined-ca-bundle\") pod \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.000299 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-nova-metadata-tls-certs\") pod \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.000373 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzvqk\" (UniqueName: \"kubernetes.io/projected/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-kube-api-access-jzvqk\") pod \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.000459 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-logs\") pod \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.000590 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-config-data\") pod \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\" (UID: \"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4\") " Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.001268 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-logs" (OuterVolumeSpecName: "logs") pod "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" (UID: "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.008583 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-kube-api-access-jzvqk" (OuterVolumeSpecName: "kube-api-access-jzvqk") pod "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" (UID: "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4"). InnerVolumeSpecName "kube-api-access-jzvqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.041508 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-config-data" (OuterVolumeSpecName: "config-data") pod "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" (UID: "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.052697 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" (UID: "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.064960 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" (UID: "5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.092602 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d18b4c1-d423-4840-97ff-c322272c3aa3","Type":"ContainerStarted","Data":"e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb"} Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.098086 4741 generic.go:334] "Generic (PLEG): container finished" podID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerID="730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6" exitCode=0 Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.098126 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.098156 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4","Type":"ContainerDied","Data":"730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6"} Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.098232 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4","Type":"ContainerDied","Data":"c20bdc770ea5d8ae7d8bfbb83c8a9b7be3a18527b638c9b18c87373d0b652d2f"} Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.098254 4741 scope.go:117] "RemoveContainer" containerID="730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.103518 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.103552 4741 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.103566 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzvqk\" (UniqueName: \"kubernetes.io/projected/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-kube-api-access-jzvqk\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.103581 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.103619 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.119081 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.119062079 podStartE2EDuration="2.119062079s" podCreationTimestamp="2025-09-29 19:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:30:20.113561897 +0000 UTC m=+1261.761351239" watchObservedRunningTime="2025-09-29 19:30:20.119062079 +0000 UTC m=+1261.766851421" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.135847 4741 scope.go:117] "RemoveContainer" containerID="8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.140139 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.149762 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.164960 4741 scope.go:117] "RemoveContainer" containerID="730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6" Sep 29 19:30:20 crc kubenswrapper[4741]: E0929 19:30:20.165662 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6\": container with ID starting with 730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6 not found: ID does not exist" containerID="730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.165713 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6"} err="failed to get container status \"730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6\": rpc error: code = NotFound desc = could not find container \"730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6\": container with ID starting with 730f4a0a19baa38a9d89aeaa3189a209df37f5d41cf0a2dae39c1111d4af36e6 not found: ID does not exist" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.165741 4741 scope.go:117] "RemoveContainer" containerID="8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1" Sep 29 19:30:20 crc kubenswrapper[4741]: E0929 19:30:20.166054 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1\": container with ID starting with 8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1 not found: ID does not exist" containerID="8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.166081 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1"} err="failed to get container status \"8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1\": rpc error: code = NotFound desc = could not find container \"8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1\": container with ID starting with 8f5fa4836263effeb2a6dc58db7cccba32c86d8b313cb57e6e331394c0ea44c1 not found: ID does not exist" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.169007 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:30:20 crc kubenswrapper[4741]: E0929 19:30:20.169486 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-metadata" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.169510 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-metadata" Sep 29 19:30:20 crc kubenswrapper[4741]: E0929 19:30:20.169531 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-log" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.169540 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-log" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.169772 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-log" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.169791 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" containerName="nova-metadata-metadata" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.172499 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.175909 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.177688 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.206320 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.311031 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b0a22c3-d524-422d-8692-14d7a16a418f-logs\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.311113 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.311249 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-config-data\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.311270 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2rwz\" (UniqueName: \"kubernetes.io/projected/6b0a22c3-d524-422d-8692-14d7a16a418f-kube-api-access-l2rwz\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.311309 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.413052 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.413364 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b0a22c3-d524-422d-8692-14d7a16a418f-logs\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.413482 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.413596 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-config-data\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.413623 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2rwz\" (UniqueName: \"kubernetes.io/projected/6b0a22c3-d524-422d-8692-14d7a16a418f-kube-api-access-l2rwz\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.414008 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b0a22c3-d524-422d-8692-14d7a16a418f-logs\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.417087 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.419274 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-config-data\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.425000 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.430256 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2rwz\" (UniqueName: \"kubernetes.io/projected/6b0a22c3-d524-422d-8692-14d7a16a418f-kube-api-access-l2rwz\") pod \"nova-metadata-0\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.497556 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:30:20 crc kubenswrapper[4741]: I0929 19:30:20.997939 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:30:21 crc kubenswrapper[4741]: I0929 19:30:21.104200 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4" path="/var/lib/kubelet/pods/5cae00e0-9012-4f8b-96fd-ab24cd1b8ad4/volumes" Sep 29 19:30:21 crc kubenswrapper[4741]: I0929 19:30:21.114487 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b0a22c3-d524-422d-8692-14d7a16a418f","Type":"ContainerStarted","Data":"376e5a13a7a12be380fb1f9484cf4feabbedc5762e3ae2e1369aadd1393f5203"} Sep 29 19:30:22 crc kubenswrapper[4741]: I0929 19:30:22.128833 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b0a22c3-d524-422d-8692-14d7a16a418f","Type":"ContainerStarted","Data":"e3ba9d353708ea132687e9ee72cfd9472209753189bc1863b2bf80c6d632fdd9"} Sep 29 19:30:22 crc kubenswrapper[4741]: I0929 19:30:22.129123 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b0a22c3-d524-422d-8692-14d7a16a418f","Type":"ContainerStarted","Data":"254ddac676d3d1212195f5f14061933f9ab70fa1f2be993950d36458f551f9d7"} Sep 29 19:30:23 crc kubenswrapper[4741]: I0929 19:30:23.434040 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 19:30:25 crc kubenswrapper[4741]: I0929 19:30:25.498626 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:30:25 crc kubenswrapper[4741]: I0929 19:30:25.499057 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 19:30:27 crc kubenswrapper[4741]: I0929 19:30:27.408516 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 19:30:27 crc kubenswrapper[4741]: I0929 19:30:27.408838 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 19:30:28 crc kubenswrapper[4741]: I0929 19:30:28.420506 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 19:30:28 crc kubenswrapper[4741]: I0929 19:30:28.420554 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 19:30:28 crc kubenswrapper[4741]: I0929 19:30:28.433416 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 19:30:28 crc kubenswrapper[4741]: I0929 19:30:28.461318 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 19:30:28 crc kubenswrapper[4741]: I0929 19:30:28.487649 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=8.487633426 podStartE2EDuration="8.487633426s" podCreationTimestamp="2025-09-29 19:30:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 19:30:22.146806628 +0000 UTC m=+1263.794595980" watchObservedRunningTime="2025-09-29 19:30:28.487633426 +0000 UTC m=+1270.135422758" Sep 29 19:30:29 crc kubenswrapper[4741]: I0929 19:30:29.238355 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 19:30:30 crc kubenswrapper[4741]: I0929 19:30:30.498595 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 19:30:30 crc kubenswrapper[4741]: I0929 19:30:30.498922 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 19:30:31 crc kubenswrapper[4741]: I0929 19:30:31.507667 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 19:30:31 crc kubenswrapper[4741]: I0929 19:30:31.515745 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 19:30:35 crc kubenswrapper[4741]: I0929 19:30:35.505740 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 29 19:30:37 crc kubenswrapper[4741]: I0929 19:30:37.416419 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 19:30:37 crc kubenswrapper[4741]: I0929 19:30:37.417072 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 19:30:37 crc kubenswrapper[4741]: I0929 19:30:37.418007 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 19:30:37 crc kubenswrapper[4741]: I0929 19:30:37.423944 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 19:30:38 crc kubenswrapper[4741]: I0929 19:30:38.307583 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 19:30:38 crc kubenswrapper[4741]: I0929 19:30:38.314654 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 19:30:40 crc kubenswrapper[4741]: I0929 19:30:40.503834 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 19:30:40 crc kubenswrapper[4741]: I0929 19:30:40.506986 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 19:30:40 crc kubenswrapper[4741]: I0929 19:30:40.508338 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 19:30:41 crc kubenswrapper[4741]: I0929 19:30:41.355625 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.690272 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.690963 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="fc0eaaf0-97ff-47ea-a15d-0024fbc96439" containerName="openstackclient" containerID="cri-o://a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead" gracePeriod=2 Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.716888 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.866237 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.923539 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.923749 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="ovn-northd" containerID="cri-o://3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96" gracePeriod=30 Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.923857 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="openstack-network-exporter" containerID="cri-o://e2974921beee8fdb66bfc522f1bf2e042117a87afcc7cd9f689c79ab47e6bcf2" gracePeriod=30 Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.956635 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement1b19-account-delete-shzs7"] Sep 29 19:31:02 crc kubenswrapper[4741]: E0929 19:31:02.957628 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0eaaf0-97ff-47ea-a15d-0024fbc96439" containerName="openstackclient" Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.957750 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0eaaf0-97ff-47ea-a15d-0024fbc96439" containerName="openstackclient" Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.958015 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0eaaf0-97ff-47ea-a15d-0024fbc96439" containerName="openstackclient" Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.958896 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement1b19-account-delete-shzs7" Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.994085 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="ovn-northd" probeResult="failure" output=< Sep 29 19:31:02 crc kubenswrapper[4741]: 2025-09-29T19:31:02Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Sep 29 19:31:02 crc kubenswrapper[4741]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Sep 29 19:31:02 crc kubenswrapper[4741]: 2025-09-29T19:31:02Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Sep 29 19:31:02 crc kubenswrapper[4741]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Sep 29 19:31:02 crc kubenswrapper[4741]: > Sep 29 19:31:02 crc kubenswrapper[4741]: I0929 19:31:02.996960 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.039044 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron07da-account-delete-vj777"] Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.070440 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.070510 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data podName:8e0c02dc-69a9-4e60-b179-0e23842d10a4 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:03.570491651 +0000 UTC m=+1305.218280983 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data") pod "rabbitmq-cell1-server-0" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4") : configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.071954 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.099718 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="ovn-northd" probeResult="failure" output=< Sep 29 19:31:03 crc kubenswrapper[4741]: 2025-09-29T19:31:03Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Sep 29 19:31:03 crc kubenswrapper[4741]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Sep 29 19:31:03 crc kubenswrapper[4741]: > Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.150404 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron07da-account-delete-vj777"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.150441 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement1b19-account-delete-shzs7"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.171839 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjw59\" (UniqueName: \"kubernetes.io/projected/b8c0591c-26f3-48eb-9553-e61c932fda9d-kube-api-access-wjw59\") pod \"placement1b19-account-delete-shzs7\" (UID: \"b8c0591c-26f3-48eb-9553-e61c932fda9d\") " pod="openstack/placement1b19-account-delete-shzs7" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.171931 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwdc8\" (UniqueName: \"kubernetes.io/projected/2094977c-9460-4dfd-926d-aac495e9bb73-kube-api-access-xwdc8\") pod \"neutron07da-account-delete-vj777\" (UID: \"2094977c-9460-4dfd-926d-aac495e9bb73\") " pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.172509 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.172554 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data podName:df6beb49-03ad-47ef-a9c7-3f37baa6d105 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:03.67254132 +0000 UTC m=+1305.320330652 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data") pod "rabbitmq-server-0" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105") : configmap "rabbitmq-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.176742 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance6af0-account-delete-8hrbb"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.181417 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance6af0-account-delete-8hrbb" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.190745 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance6af0-account-delete-8hrbb"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.222901 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hjb7q"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.233766 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hjb7q"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.274643 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-564rb\" (UniqueName: \"kubernetes.io/projected/983d2897-24d2-462c-b9f9-427639d2b8f9-kube-api-access-564rb\") pod \"glance6af0-account-delete-8hrbb\" (UID: \"983d2897-24d2-462c-b9f9-427639d2b8f9\") " pod="openstack/glance6af0-account-delete-8hrbb" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.274784 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjw59\" (UniqueName: \"kubernetes.io/projected/b8c0591c-26f3-48eb-9553-e61c932fda9d-kube-api-access-wjw59\") pod \"placement1b19-account-delete-shzs7\" (UID: \"b8c0591c-26f3-48eb-9553-e61c932fda9d\") " pod="openstack/placement1b19-account-delete-shzs7" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.274878 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwdc8\" (UniqueName: \"kubernetes.io/projected/2094977c-9460-4dfd-926d-aac495e9bb73-kube-api-access-xwdc8\") pod \"neutron07da-account-delete-vj777\" (UID: \"2094977c-9460-4dfd-926d-aac495e9bb73\") " pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.301954 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.302505 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="openstack-network-exporter" containerID="cri-o://af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188" gracePeriod=300 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.307150 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwdc8\" (UniqueName: \"kubernetes.io/projected/2094977c-9460-4dfd-926d-aac495e9bb73-kube-api-access-xwdc8\") pod \"neutron07da-account-delete-vj777\" (UID: \"2094977c-9460-4dfd-926d-aac495e9bb73\") " pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.314214 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjw59\" (UniqueName: \"kubernetes.io/projected/b8c0591c-26f3-48eb-9553-e61c932fda9d-kube-api-access-wjw59\") pod \"placement1b19-account-delete-shzs7\" (UID: \"b8c0591c-26f3-48eb-9553-e61c932fda9d\") " pod="openstack/placement1b19-account-delete-shzs7" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.340294 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement1b19-account-delete-shzs7" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.347183 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-xlz27"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.400932 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-xlz27"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.482672 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="ovsdbserver-nb" containerID="cri-o://c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f" gracePeriod=300 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.511448 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-564rb\" (UniqueName: \"kubernetes.io/projected/983d2897-24d2-462c-b9f9-427639d2b8f9-kube-api-access-564rb\") pod \"glance6af0-account-delete-8hrbb\" (UID: \"983d2897-24d2-462c-b9f9-427639d2b8f9\") " pod="openstack/glance6af0-account-delete-8hrbb" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.528023 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.547772 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-564rb\" (UniqueName: \"kubernetes.io/projected/983d2897-24d2-462c-b9f9-427639d2b8f9-kube-api-access-564rb\") pod \"glance6af0-account-delete-8hrbb\" (UID: \"983d2897-24d2-462c-b9f9-427639d2b8f9\") " pod="openstack/glance6af0-account-delete-8hrbb" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.559652 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell15b24-account-delete-whtzz"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.572660 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell15b24-account-delete-whtzz"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.572871 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell15b24-account-delete-whtzz" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.591641 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi7d7d-account-delete-wklvh"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.592903 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi7d7d-account-delete-wklvh" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.604643 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi7d7d-account-delete-wklvh"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.608695 4741 generic.go:334] "Generic (PLEG): container finished" podID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerID="e2974921beee8fdb66bfc522f1bf2e042117a87afcc7cd9f689c79ab47e6bcf2" exitCode=2 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.608751 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9d4bbc33-c647-4eda-bdc6-89f80678ae8d","Type":"ContainerDied","Data":"e2974921beee8fdb66bfc522f1bf2e042117a87afcc7cd9f689c79ab47e6bcf2"} Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.614721 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.614795 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data podName:8e0c02dc-69a9-4e60-b179-0e23842d10a4 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:04.614776655 +0000 UTC m=+1306.262565987 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data") pod "rabbitmq-cell1-server-0" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4") : configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.617446 4741 generic.go:334] "Generic (PLEG): container finished" podID="f5597514-2fd8-4d92-b115-05b66894ea94" containerID="af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188" exitCode=2 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.617470 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f5597514-2fd8-4d92-b115-05b66894ea94","Type":"ContainerDied","Data":"af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188"} Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.691997 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.692916 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="openstack-network-exporter" containerID="cri-o://b10e00d6e3d0b86eee6df15910497c173193cb10140a8e44a1e3e58aa399a2cd" gracePeriod=300 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.705436 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.719490 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnhdk\" (UniqueName: \"kubernetes.io/projected/9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62-kube-api-access-xnhdk\") pod \"novaapi7d7d-account-delete-wklvh\" (UID: \"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62\") " pod="openstack/novaapi7d7d-account-delete-wklvh" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.719552 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bjgr\" (UniqueName: \"kubernetes.io/projected/13785b5b-f629-47ae-8251-331d44b40254-kube-api-access-7bjgr\") pod \"novacell15b24-account-delete-whtzz\" (UID: \"13785b5b-f629-47ae-8251-331d44b40254\") " pod="openstack/novacell15b24-account-delete-whtzz" Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.719674 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: E0929 19:31:03.719711 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data podName:df6beb49-03ad-47ef-a9c7-3f37baa6d105 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:04.719697304 +0000 UTC m=+1306.367486636 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data") pod "rabbitmq-server-0" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105") : configmap "rabbitmq-config-data" not found Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.744456 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-z95rt"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.765959 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-z95rt"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.770486 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerName="rabbitmq" containerID="cri-o://7d7103d33c61d8d8489afa8a3b1ea5db27133d9afa53b11454cfade2f1762a5c" gracePeriod=604800 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.813253 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance6af0-account-delete-8hrbb" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.822585 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnhdk\" (UniqueName: \"kubernetes.io/projected/9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62-kube-api-access-xnhdk\") pod \"novaapi7d7d-account-delete-wklvh\" (UID: \"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62\") " pod="openstack/novaapi7d7d-account-delete-wklvh" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.822662 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bjgr\" (UniqueName: \"kubernetes.io/projected/13785b5b-f629-47ae-8251-331d44b40254-kube-api-access-7bjgr\") pod \"novacell15b24-account-delete-whtzz\" (UID: \"13785b5b-f629-47ae-8251-331d44b40254\") " pod="openstack/novacell15b24-account-delete-whtzz" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.838681 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="ovsdbserver-sb" containerID="cri-o://814a325d7b6c7b6ac7fd7b9d3bac9dffa54a1007b3dabf568b1a0a21bc488445" gracePeriod=300 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.850227 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bjgr\" (UniqueName: \"kubernetes.io/projected/13785b5b-f629-47ae-8251-331d44b40254-kube-api-access-7bjgr\") pod \"novacell15b24-account-delete-whtzz\" (UID: \"13785b5b-f629-47ae-8251-331d44b40254\") " pod="openstack/novacell15b24-account-delete-whtzz" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.850824 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnhdk\" (UniqueName: \"kubernetes.io/projected/9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62-kube-api-access-xnhdk\") pod \"novaapi7d7d-account-delete-wklvh\" (UID: \"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62\") " pod="openstack/novaapi7d7d-account-delete-wklvh" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.909856 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.961230 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell15b24-account-delete-whtzz" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.980306 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerName="rabbitmq" containerID="cri-o://05c981478a21299cdffcc57230be7582dcc6e046d2f10f197e1a1fb4606f05e6" gracePeriod=604800 Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.989671 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi7d7d-account-delete-wklvh" Sep 29 19:31:03 crc kubenswrapper[4741]: I0929 19:31:03.990563 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-68snt" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerName="ovn-controller" probeResult="failure" output=< Sep 29 19:31:03 crc kubenswrapper[4741]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 29 19:31:03 crc kubenswrapper[4741]: > Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.081552 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-646p8"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.090599 4741 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron-5dd77b8d5c-hw4v7" secret="" err="secret \"neutron-neutron-dockercfg-c9t62\" not found" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.120449 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-646p8"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.144160 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-gfpz8"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.161779 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-gfpz8"] Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.246062 4741 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.246151 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:04.746132307 +0000 UTC m=+1306.393921639 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.252427 4741 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.252514 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:04.752496087 +0000 UTC m=+1306.400285419 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-httpd-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.311446 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c686d89f8-xwpjv"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.311678 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-c686d89f8-xwpjv" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-log" containerID="cri-o://c7249b3b7837fc3520fcc2d531958323f2ab743d34b60e50c851a65328623db0" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.312088 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-c686d89f8-xwpjv" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-api" containerID="cri-o://88822bc2858601e3da0fba4f6a6ca50dcd1d5c1fe288573a193d9458b638d155" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.351671 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f5597514-2fd8-4d92-b115-05b66894ea94/ovsdbserver-nb/0.log" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.358788 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.375176 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-q65b6"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.398205 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fkpb5"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.398763 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-fkpb5" podUID="3c11cbdf-1417-4c8d-b3bb-150f57496287" containerName="openstack-network-exporter" containerID="cri-o://ae21046b33ac269496303f0c89312707199e95ccc69f985f150deaee38b8c665" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.420682 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-68snt"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.440361 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-hbdwb"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.447354 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-hbdwb"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.457667 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dbj47"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.469999 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.501508 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fmrvg"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.522026 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dbj47"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.542817 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fmrvg"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.557153 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-metrics-certs-tls-certs\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.557261 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-config\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.557317 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-scripts\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.558099 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-config" (OuterVolumeSpecName: "config") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.558212 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-scripts" (OuterVolumeSpecName: "scripts") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.558342 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdb-rundir\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.558373 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhxsp\" (UniqueName: \"kubernetes.io/projected/f5597514-2fd8-4d92-b115-05b66894ea94-kube-api-access-dhxsp\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.558899 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdbserver-nb-tls-certs\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.559012 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-combined-ca-bundle\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.559138 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"f5597514-2fd8-4d92-b115-05b66894ea94\" (UID: \"f5597514-2fd8-4d92-b115-05b66894ea94\") " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.559898 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.559919 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5597514-2fd8-4d92-b115-05b66894ea94-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.563783 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.564313 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dd77b8d5c-hw4v7"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.565031 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.567944 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5597514-2fd8-4d92-b115-05b66894ea94-kube-api-access-dhxsp" (OuterVolumeSpecName: "kube-api-access-dhxsp") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "kube-api-access-dhxsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.601522 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-55hm7"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.601775 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerName="dnsmasq-dns" containerID="cri-o://1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da" gracePeriod=10 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.619298 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.619632 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.624762 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-server" containerID="cri-o://0f18bf31b15ac6485b4321250e71d201fadbe48a440243ff5584feadda173ffc" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.624890 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="swift-recon-cron" containerID="cri-o://e8277ed6dfec4fbba8595886bec52f0c929d6c4a577686a175df2b460882c90f" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.624926 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="rsync" containerID="cri-o://af59f98d013533f7b1f00c005e49813bc7d0e7ab704e34a34db4cec3151d2b45" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.624964 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-expirer" containerID="cri-o://a979b22c48d4833060b798e59c6172ae0d0fb020aa7081564218be678d051ef8" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.624996 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-updater" containerID="cri-o://878b7f5b4bcf11108cbc1b12fcff16a252c21bb2290b20babdfe2cc60f5f4278" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625026 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-auditor" containerID="cri-o://a68bac92910f13da6b5bd1c4c6fa29f7fec8ba0ea712f3e3c9da7813f47eafee" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625057 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-replicator" containerID="cri-o://5ace69d8ba764865fefb3a17c320882c81b7b6f3fba4329e49c21e37a73ff7f1" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625090 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-server" containerID="cri-o://71e20963cd9fad0ee127b4dda01c84dd767efbefb2fcb80296923f8718f1ddbf" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625122 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-updater" containerID="cri-o://49334c9dcc9f0ab359baa344e3fe68ec6f0e5271440eec25dd046555f7cc0e20" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625165 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-auditor" containerID="cri-o://8bdeba0cf6a202fe00074a937a1baa94541c1fc21049258f466a2ee65af3a75e" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625194 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-replicator" containerID="cri-o://e650c5ffc6ebc6918dc63202d7b9d6aaff80d4b6db1471fd84f3faebfd8b934f" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625226 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-server" containerID="cri-o://0272c3b7736bc9733f76397bbd76a46beab031e046ac627b3f5b05578332ba13" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625278 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-reaper" containerID="cri-o://11c745f1505ad37096a8441c7db5e1f99996efc6719e0f171773f7647d65d6f0" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625319 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-auditor" containerID="cri-o://5c99d27cbf271df86190fc67b6bd3129d7c9ebc403240020a83b932781c09aa8" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625351 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-replicator" containerID="cri-o://bdcc643cf7b0ba52293b398520b054d5a0af7419de83697c00319b14e82a22f5" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.625883 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.626059 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-log" containerID="cri-o://59e07c173f297fc52b58154fee10e2757e48b2c11cd92731bb4e2fc98f386bc7" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.626138 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-httpd" containerID="cri-o://b426eb53e2825c670095e43996bf58ecec4760a4bafd71a4d7304962625a63c7" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.631613 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.632257 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-log" containerID="cri-o://e1b73040b554cbd478a51c8b5816e58a0a68a53c90c12ff3ca2f8c2045c814bf" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.632421 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-httpd" containerID="cri-o://c9909a6976e00256c9ff763bed1a932c9df0fd365ff4dabc64990d63e5be6cf4" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.685807 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.686059 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.686154 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhxsp\" (UniqueName: \"kubernetes.io/projected/f5597514-2fd8-4d92-b115-05b66894ea94-kube-api-access-dhxsp\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.686292 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.686176 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.686487 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data podName:8e0c02dc-69a9-4e60-b179-0e23842d10a4 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:06.686469733 +0000 UTC m=+1308.334259065 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data") pod "rabbitmq-cell1-server-0" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4") : configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.697192 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f5597514-2fd8-4d92-b115-05b66894ea94/ovsdbserver-nb/0.log" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.703482 4741 generic.go:334] "Generic (PLEG): container finished" podID="f5597514-2fd8-4d92-b115-05b66894ea94" containerID="c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f" exitCode=143 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.703599 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f5597514-2fd8-4d92-b115-05b66894ea94","Type":"ContainerDied","Data":"c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f"} Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.703631 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f5597514-2fd8-4d92-b115-05b66894ea94","Type":"ContainerDied","Data":"89bf1a08d6d1d7b35f54b96a50f9eb7fb63f0f2b6966359c9c14f1093319b854"} Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.703657 4741 scope.go:117] "RemoveContainer" containerID="c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.703829 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.764572 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.764890 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="cinder-scheduler" containerID="cri-o://a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.765182 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="probe" containerID="cri-o://5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.775076 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_69f1c2bd-c091-42e8-8810-a04726ce9032/ovsdbserver-sb/0.log" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.775117 4741 generic.go:334] "Generic (PLEG): container finished" podID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerID="b10e00d6e3d0b86eee6df15910497c173193cb10140a8e44a1e3e58aa399a2cd" exitCode=2 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.775157 4741 generic.go:334] "Generic (PLEG): container finished" podID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerID="814a325d7b6c7b6ac7fd7b9d3bac9dffa54a1007b3dabf568b1a0a21bc488445" exitCode=143 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.775243 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69f1c2bd-c091-42e8-8810-a04726ce9032","Type":"ContainerDied","Data":"b10e00d6e3d0b86eee6df15910497c173193cb10140a8e44a1e3e58aa399a2cd"} Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.775282 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69f1c2bd-c091-42e8-8810-a04726ce9032","Type":"ContainerDied","Data":"814a325d7b6c7b6ac7fd7b9d3bac9dffa54a1007b3dabf568b1a0a21bc488445"} Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.796516 4741 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.796618 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:05.796588446 +0000 UTC m=+1307.444377778 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.797430 4741 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.797523 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:05.797503814 +0000 UTC m=+1307.445293146 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-httpd-config" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.797917 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Sep 29 19:31:04 crc kubenswrapper[4741]: E0929 19:31:04.798121 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data podName:df6beb49-03ad-47ef-a9c7-3f37baa6d105 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:06.798108883 +0000 UTC m=+1308.445898215 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data") pod "rabbitmq-server-0" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105") : configmap "rabbitmq-config-data" not found Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.800500 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fkpb5_3c11cbdf-1417-4c8d-b3bb-150f57496287/openstack-network-exporter/0.log" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.800563 4741 generic.go:334] "Generic (PLEG): container finished" podID="3c11cbdf-1417-4c8d-b3bb-150f57496287" containerID="ae21046b33ac269496303f0c89312707199e95ccc69f985f150deaee38b8c665" exitCode=2 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.800650 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fkpb5" event={"ID":"3c11cbdf-1417-4c8d-b3bb-150f57496287","Type":"ContainerDied","Data":"ae21046b33ac269496303f0c89312707199e95ccc69f985f150deaee38b8c665"} Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.874088 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.889679 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.896131 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-fqvpl"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.896458 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.898696 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "f5597514-2fd8-4d92-b115-05b66894ea94" (UID: "f5597514-2fd8-4d92-b115-05b66894ea94"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.900639 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.900660 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.900670 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5597514-2fd8-4d92-b115-05b66894ea94-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.908438 4741 generic.go:334] "Generic (PLEG): container finished" podID="2be2676f-d882-424e-8a49-d8bffc23773e" containerID="c7249b3b7837fc3520fcc2d531958323f2ab743d34b60e50c851a65328623db0" exitCode=143 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.908675 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dd77b8d5c-hw4v7" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-api" containerID="cri-o://3d283ad6e8c834fde894dcf133b306ef1602160587b86f10edad93244d544e8f" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.909083 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c686d89f8-xwpjv" event={"ID":"2be2676f-d882-424e-8a49-d8bffc23773e","Type":"ContainerDied","Data":"c7249b3b7837fc3520fcc2d531958323f2ab743d34b60e50c851a65328623db0"} Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.909141 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dd77b8d5c-hw4v7" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-httpd" containerID="cri-o://1700c1e978f1aa2ccdb75b0383b4cb4e5f89351b953d72fcb8ea36e16ead7085" gracePeriod=30 Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.934685 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1b19-account-create-clkpq"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.948174 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-fqvpl"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.966835 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1b19-account-create-clkpq"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.984035 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement1b19-account-delete-shzs7"] Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.986027 4741 scope.go:117] "RemoveContainer" containerID="af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188" Sep 29 19:31:04 crc kubenswrapper[4741]: I0929 19:31:04.989379 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-2s8qk"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.003579 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-2s8qk"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.014426 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.014633 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api-log" containerID="cri-o://eac6f805571d33e47abcb8386b1e4a78317268ba124fe10a2e9376e94d5dd506" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.014977 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api" containerID="cri-o://5747543db2d4686ff0ce2d66484ba05733793d16bc7f1a3c6bf312799d425845" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.027646 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron07da-account-delete-vj777"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.033152 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_69f1c2bd-c091-42e8-8810-a04726ce9032/ovsdbserver-sb/0.log" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.033225 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.044809 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-mrkv4"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.058954 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-mrkv4"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.063343 4741 scope.go:117] "RemoveContainer" containerID="c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f" Sep 29 19:31:05 crc kubenswrapper[4741]: E0929 19:31:05.066507 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f\": container with ID starting with c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f not found: ID does not exist" containerID="c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.066544 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f"} err="failed to get container status \"c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f\": rpc error: code = NotFound desc = could not find container \"c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f\": container with ID starting with c4ef2d90f06647234eeeff2278c2713ce0bfbb46ac1b6959135b44a1bb3da25f not found: ID does not exist" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.066573 4741 scope.go:117] "RemoveContainer" containerID="af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188" Sep 29 19:31:05 crc kubenswrapper[4741]: E0929 19:31:05.077987 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188\": container with ID starting with af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188 not found: ID does not exist" containerID="af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.078035 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188"} err="failed to get container status \"af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188\": rpc error: code = NotFound desc = could not find container \"af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188\": container with ID starting with af163ccc60455997d20bd892bcb8b93978c22ba33fd325715aa9730eea772188 not found: ID does not exist" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.079314 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-07da-account-create-wcv7g"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.120151 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-scripts\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.120200 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnn2l\" (UniqueName: \"kubernetes.io/projected/69f1c2bd-c091-42e8-8810-a04726ce9032-kube-api-access-mnn2l\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.120254 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-metrics-certs-tls-certs\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.120275 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdbserver-sb-tls-certs\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.120900 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-combined-ca-bundle\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.121013 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.121057 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-config\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.121911 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdb-rundir\") pod \"69f1c2bd-c091-42e8-8810-a04726ce9032\" (UID: \"69f1c2bd-c091-42e8-8810-a04726ce9032\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.132170 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f1c2bd-c091-42e8-8810-a04726ce9032-kube-api-access-mnn2l" (OuterVolumeSpecName: "kube-api-access-mnn2l") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "kube-api-access-mnn2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.132958 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnn2l\" (UniqueName: \"kubernetes.io/projected/69f1c2bd-c091-42e8-8810-a04726ce9032-kube-api-access-mnn2l\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.134443 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.135712 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-scripts" (OuterVolumeSpecName: "scripts") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.135762 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" containerID="cri-o://e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.138945 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-config" (OuterVolumeSpecName: "config") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.165696 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: E0929 19:31:05.166723 4741 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Sep 29 19:31:05 crc kubenswrapper[4741]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Sep 29 19:31:05 crc kubenswrapper[4741]: + source /usr/local/bin/container-scripts/functions Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNBridge=br-int Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNRemote=tcp:localhost:6642 Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNEncapType=geneve Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNAvailabilityZones= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ EnableChassisAsGateway=true Sep 29 19:31:05 crc kubenswrapper[4741]: ++ PhysicalNetworks= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNHostName= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ DB_FILE=/etc/openvswitch/conf.db Sep 29 19:31:05 crc kubenswrapper[4741]: ++ ovs_dir=/var/lib/openvswitch Sep 29 19:31:05 crc kubenswrapper[4741]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Sep 29 19:31:05 crc kubenswrapper[4741]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Sep 29 19:31:05 crc kubenswrapper[4741]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Sep 29 19:31:05 crc kubenswrapper[4741]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Sep 29 19:31:05 crc kubenswrapper[4741]: + sleep 0.5 Sep 29 19:31:05 crc kubenswrapper[4741]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Sep 29 19:31:05 crc kubenswrapper[4741]: + cleanup_ovsdb_server_semaphore Sep 29 19:31:05 crc kubenswrapper[4741]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Sep 29 19:31:05 crc kubenswrapper[4741]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Sep 29 19:31:05 crc kubenswrapper[4741]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-q65b6" message=< Sep 29 19:31:05 crc kubenswrapper[4741]: Exiting ovsdb-server (5) [ OK ] Sep 29 19:31:05 crc kubenswrapper[4741]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Sep 29 19:31:05 crc kubenswrapper[4741]: + source /usr/local/bin/container-scripts/functions Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNBridge=br-int Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNRemote=tcp:localhost:6642 Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNEncapType=geneve Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNAvailabilityZones= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ EnableChassisAsGateway=true Sep 29 19:31:05 crc kubenswrapper[4741]: ++ PhysicalNetworks= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNHostName= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ DB_FILE=/etc/openvswitch/conf.db Sep 29 19:31:05 crc kubenswrapper[4741]: ++ ovs_dir=/var/lib/openvswitch Sep 29 19:31:05 crc kubenswrapper[4741]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Sep 29 19:31:05 crc kubenswrapper[4741]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Sep 29 19:31:05 crc kubenswrapper[4741]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Sep 29 19:31:05 crc kubenswrapper[4741]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Sep 29 19:31:05 crc kubenswrapper[4741]: + sleep 0.5 Sep 29 19:31:05 crc kubenswrapper[4741]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Sep 29 19:31:05 crc kubenswrapper[4741]: + cleanup_ovsdb_server_semaphore Sep 29 19:31:05 crc kubenswrapper[4741]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Sep 29 19:31:05 crc kubenswrapper[4741]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Sep 29 19:31:05 crc kubenswrapper[4741]: > Sep 29 19:31:05 crc kubenswrapper[4741]: E0929 19:31:05.166771 4741 kuberuntime_container.go:691] "PreStop hook failed" err=< Sep 29 19:31:05 crc kubenswrapper[4741]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Sep 29 19:31:05 crc kubenswrapper[4741]: + source /usr/local/bin/container-scripts/functions Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNBridge=br-int Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNRemote=tcp:localhost:6642 Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNEncapType=geneve Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNAvailabilityZones= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ EnableChassisAsGateway=true Sep 29 19:31:05 crc kubenswrapper[4741]: ++ PhysicalNetworks= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ OVNHostName= Sep 29 19:31:05 crc kubenswrapper[4741]: ++ DB_FILE=/etc/openvswitch/conf.db Sep 29 19:31:05 crc kubenswrapper[4741]: ++ ovs_dir=/var/lib/openvswitch Sep 29 19:31:05 crc kubenswrapper[4741]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Sep 29 19:31:05 crc kubenswrapper[4741]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Sep 29 19:31:05 crc kubenswrapper[4741]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Sep 29 19:31:05 crc kubenswrapper[4741]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Sep 29 19:31:05 crc kubenswrapper[4741]: + sleep 0.5 Sep 29 19:31:05 crc kubenswrapper[4741]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Sep 29 19:31:05 crc kubenswrapper[4741]: + cleanup_ovsdb_server_semaphore Sep 29 19:31:05 crc kubenswrapper[4741]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Sep 29 19:31:05 crc kubenswrapper[4741]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Sep 29 19:31:05 crc kubenswrapper[4741]: > pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" containerID="cri-o://f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.166812 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" containerID="cri-o://f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.220747 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="023a246b-31d3-47bb-822b-c8e61aa2f034" path="/var/lib/kubelet/pods/023a246b-31d3-47bb-822b-c8e61aa2f034/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.221464 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="029559ea-d2e3-4a1b-87b8-60b0a65c9730" path="/var/lib/kubelet/pods/029559ea-d2e3-4a1b-87b8-60b0a65c9730/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.222039 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13ca2f00-5599-4912-83dd-afbe369673b6" path="/var/lib/kubelet/pods/13ca2f00-5599-4912-83dd-afbe369673b6/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.224059 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e91d309-4c23-4d57-ad3f-8eb7634b9b04" path="/var/lib/kubelet/pods/2e91d309-4c23-4d57-ad3f-8eb7634b9b04/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.224567 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a7cc159-83d9-406f-8e58-23795673b6b0" path="/var/lib/kubelet/pods/4a7cc159-83d9-406f-8e58-23795673b6b0/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.225143 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629809b6-c9b1-4d39-a89c-4d8418e1bdc0" path="/var/lib/kubelet/pods/629809b6-c9b1-4d39-a89c-4d8418e1bdc0/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.225662 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e81d61-d517-412d-a5cb-100a5aac1ec3" path="/var/lib/kubelet/pods/72e81d61-d517-412d-a5cb-100a5aac1ec3/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.230371 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acf52497-97c7-4e1f-a79c-29ec872359c5" path="/var/lib/kubelet/pods/acf52497-97c7-4e1f-a79c-29ec872359c5/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.233120 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba2de5ec-87f1-4387-befa-d853e4f877dc" path="/var/lib/kubelet/pods/ba2de5ec-87f1-4387-befa-d853e4f877dc/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.234572 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2aad5b5-c32e-4d21-b66c-2b9b39e6341e" path="/var/lib/kubelet/pods/c2aad5b5-c32e-4d21-b66c-2b9b39e6341e/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.235041 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccccac78-84c0-4d3b-a0b5-6d29323558f3" path="/var/lib/kubelet/pods/ccccac78-84c0-4d3b-a0b5-6d29323558f3/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.235215 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.235247 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.235257 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.235267 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f1c2bd-c091-42e8-8810-a04726ce9032-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.236223 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e172bee5-8543-4044-bf71-57e488d14fe4" path="/var/lib/kubelet/pods/e172bee5-8543-4044-bf71-57e488d14fe4/volumes" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.236706 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.193:5353: connect: connection refused" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.250123 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.273505 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.304924 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315687 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-07da-account-create-wcv7g"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315760 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e788-account-create-f9d6m"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315774 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e788-account-create-f9d6m"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315784 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dn5tb"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315809 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dn5tb"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315819 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6af0-account-create-6pmp5"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315828 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-hqxwh"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315837 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6af0-account-create-6pmp5"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315850 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-hqxwh"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315858 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6af0-account-delete-8hrbb"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315885 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-b41f-account-create-nsqlq"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315894 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-b41f-account-create-nsqlq"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.315904 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.316380 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.318541 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-api" containerID="cri-o://040aa5ec5b498aa5b60ba516e013b9839f5b611e8d7d5311e0be82ba0ff8b3c7" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.316164 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-log" containerID="cri-o://8dccd7363f3ef4f5fc4ea08cd195ce440540c0c08f0954b35a076493bfa65bf4" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.337034 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.337062 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.337071 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.350524 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.350791 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-log" containerID="cri-o://254ddac676d3d1212195f5f14061933f9ab70fa1f2be993950d36458f551f9d7" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.351237 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-metadata" containerID="cri-o://e3ba9d353708ea132687e9ee72cfd9472209753189bc1863b2bf80c6d632fdd9" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.351351 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fkpb5_3c11cbdf-1417-4c8d-b3bb-150f57496287/openstack-network-exporter/0.log" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.351431 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.354457 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "69f1c2bd-c091-42e8-8810-a04726ce9032" (UID: "69f1c2bd-c091-42e8-8810-a04726ce9032"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.390612 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-2nvjd"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.397759 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-2nvjd"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.407676 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2245-account-create-t9wn8"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.415995 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5ffsc"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.421073 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2245-account-create-t9wn8"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.424408 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5ffsc"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.434810 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.438042 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv5tn\" (UniqueName: \"kubernetes.io/projected/3c11cbdf-1417-4c8d-b3bb-150f57496287-kube-api-access-sv5tn\") pod \"3c11cbdf-1417-4c8d-b3bb-150f57496287\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.438109 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c11cbdf-1417-4c8d-b3bb-150f57496287-config\") pod \"3c11cbdf-1417-4c8d-b3bb-150f57496287\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.438140 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovn-rundir\") pod \"3c11cbdf-1417-4c8d-b3bb-150f57496287\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.438262 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-metrics-certs-tls-certs\") pod \"3c11cbdf-1417-4c8d-b3bb-150f57496287\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.438298 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovs-rundir\") pod \"3c11cbdf-1417-4c8d-b3bb-150f57496287\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.438345 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-combined-ca-bundle\") pod \"3c11cbdf-1417-4c8d-b3bb-150f57496287\" (UID: \"3c11cbdf-1417-4c8d-b3bb-150f57496287\") " Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.438811 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f1c2bd-c091-42e8-8810-a04726ce9032-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.448695 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "3c11cbdf-1417-4c8d-b3bb-150f57496287" (UID: "3c11cbdf-1417-4c8d-b3bb-150f57496287"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.450146 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c11cbdf-1417-4c8d-b3bb-150f57496287-config" (OuterVolumeSpecName: "config") pod "3c11cbdf-1417-4c8d-b3bb-150f57496287" (UID: "3c11cbdf-1417-4c8d-b3bb-150f57496287"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.450217 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7d7d-account-create-wwqpm"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.450266 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "3c11cbdf-1417-4c8d-b3bb-150f57496287" (UID: "3c11cbdf-1417-4c8d-b3bb-150f57496287"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.454465 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c11cbdf-1417-4c8d-b3bb-150f57496287-kube-api-access-sv5tn" (OuterVolumeSpecName: "kube-api-access-sv5tn") pod "3c11cbdf-1417-4c8d-b3bb-150f57496287" (UID: "3c11cbdf-1417-4c8d-b3bb-150f57496287"). InnerVolumeSpecName "kube-api-access-sv5tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.460426 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7d7d-account-create-wwqpm"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.468621 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi7d7d-account-delete-wklvh"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.495010 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c11cbdf-1417-4c8d-b3bb-150f57496287" (UID: "3c11cbdf-1417-4c8d-b3bb-150f57496287"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.508103 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerName="galera" containerID="cri-o://15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.511116 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5c96956bc-7qhgv"] Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.511382 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5c96956bc-7qhgv" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker-log" containerID="cri-o://f4540013c871a62255548905e7c7a5f8f143d8e4617fe18f7c6f7bdbd529b257" gracePeriod=30 Sep 29 19:31:05 crc kubenswrapper[4741]: I0929 19:31:05.511509 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5c96956bc-7qhgv" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker" containerID="cri-o://5d7c9dcdb6e6631c524f68a99886213580f717eb067525088f01ff7e8451b3e4" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.529712 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-8567fd5894-dwdtp"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.529989 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener-log" containerID="cri-o://fc163c8b456c4eae3276be79b996f9a990070e39bb81a710df286eea59af6f36" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.530163 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener" containerID="cri-o://4b30dc6f2664d0596e57ddb9855aca996f10f7caab7b0df97ee564ddcc5bc232" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.539916 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-combined-ca-bundle\") pod \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.540020 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config\") pod \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.540084 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p9s5\" (UniqueName: \"kubernetes.io/projected/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-kube-api-access-7p9s5\") pod \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.540230 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config-secret\") pod \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\" (UID: \"fc0eaaf0-97ff-47ea-a15d-0024fbc96439\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.540999 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv5tn\" (UniqueName: \"kubernetes.io/projected/3c11cbdf-1417-4c8d-b3bb-150f57496287-kube-api-access-sv5tn\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.541014 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c11cbdf-1417-4c8d-b3bb-150f57496287-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.541036 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovn-rundir\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.541055 4741 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3c11cbdf-1417-4c8d-b3bb-150f57496287-ovs-rundir\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.541064 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.544713 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-659f5d9f44-j8zgh"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.544974 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-659f5d9f44-j8zgh" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api-log" containerID="cri-o://58644df2679ff5827ca80923ddee91fde9b4da2f9eb6e37ae982ca0c9572be55" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.545118 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-659f5d9f44-j8zgh" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api" containerID="cri-o://dfd2422ee682a56f33a264c928c2202071429b72837518805278e82828695484" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.608753 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-kube-api-access-7p9s5" (OuterVolumeSpecName: "kube-api-access-7p9s5") pod "fc0eaaf0-97ff-47ea-a15d-0024fbc96439" (UID: "fc0eaaf0-97ff-47ea-a15d-0024fbc96439"). InnerVolumeSpecName "kube-api-access-7p9s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.630083 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.630332 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="f24ec6be-6832-4eee-916c-2237e81bf0f1" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.643194 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.643508 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9d18b4c1-d423-4840-97ff-c322272c3aa3" containerName="nova-scheduler-scheduler" containerID="cri-o://e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.648286 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p9s5\" (UniqueName: \"kubernetes.io/projected/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-kube-api-access-7p9s5\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.649502 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8wwcs"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.672477 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.672824 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="93c5456d-7a59-4991-ad3d-58b04db78b24" containerName="nova-cell1-conductor-conductor" containerID="cri-o://aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.726456 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8wwcs"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.744725 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.745018 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="cb846a7f-9be5-4a52-837e-a423f7736e79" containerName="nova-cell0-conductor-conductor" containerID="cri-o://e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.749790 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.754281 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b64h2"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.761990 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b64h2"] Sep 29 19:31:06 crc kubenswrapper[4741]: W0929 19:31:05.771097 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod983d2897_24d2_462c_b9f9_427639d2b8f9.slice/crio-749520f305926a8a1fb56868f7cbd0a33f054e60841c28493f2b2a74e2842488 WatchSource:0}: Error finding container 749520f305926a8a1fb56868f7cbd0a33f054e60841c28493f2b2a74e2842488: Status 404 returned error can't find the container with id 749520f305926a8a1fb56868f7cbd0a33f054e60841c28493f2b2a74e2842488 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.780571 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron07da-account-delete-vj777"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.802141 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement1b19-account-delete-shzs7"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.802332 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc0eaaf0-97ff-47ea-a15d-0024fbc96439" (UID: "fc0eaaf0-97ff-47ea-a15d-0024fbc96439"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.808408 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.817792 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.835980 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "3c11cbdf-1417-4c8d-b3bb-150f57496287" (UID: "3c11cbdf-1417-4c8d-b3bb-150f57496287"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.839595 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6af0-account-delete-8hrbb"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.846991 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell15b24-account-delete-whtzz"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.850921 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-swift-storage-0\") pod \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.851017 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-svc\") pod \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.851054 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8jc4\" (UniqueName: \"kubernetes.io/projected/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-kube-api-access-f8jc4\") pod \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.851099 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-sb\") pod \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.851144 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-nb\") pod \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.851312 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-config\") pod \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\" (UID: \"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.851858 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c11cbdf-1417-4c8d-b3bb-150f57496287-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.851873 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:05.851947 4741 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:05.851991 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:07.851977501 +0000 UTC m=+1309.499766833 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-config" not found Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.853821 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "fc0eaaf0-97ff-47ea-a15d-0024fbc96439" (UID: "fc0eaaf0-97ff-47ea-a15d-0024fbc96439"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:05.854746 4741 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:05.854807 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:07.85479243 +0000 UTC m=+1309.502581842 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-httpd-config" not found Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.870542 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "fc0eaaf0-97ff-47ea-a15d-0024fbc96439" (UID: "fc0eaaf0-97ff-47ea-a15d-0024fbc96439"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.877252 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-kube-api-access-f8jc4" (OuterVolumeSpecName: "kube-api-access-f8jc4") pod "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" (UID: "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb"). InnerVolumeSpecName "kube-api-access-f8jc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.937163 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi7d7d-account-delete-wklvh"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.943465 4741 generic.go:334] "Generic (PLEG): container finished" podID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerID="1700c1e978f1aa2ccdb75b0383b4cb4e5f89351b953d72fcb8ea36e16ead7085" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.943539 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd77b8d5c-hw4v7" event={"ID":"760672f8-7a2f-4ddb-8ebd-3664670a4adc","Type":"ContainerDied","Data":"1700c1e978f1aa2ccdb75b0383b4cb4e5f89351b953d72fcb8ea36e16ead7085"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.945309 4741 generic.go:334] "Generic (PLEG): container finished" podID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerID="254ddac676d3d1212195f5f14061933f9ab70fa1f2be993950d36458f551f9d7" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.945347 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b0a22c3-d524-422d-8692-14d7a16a418f","Type":"ContainerDied","Data":"254ddac676d3d1212195f5f14061933f9ab70fa1f2be993950d36458f551f9d7"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.947137 4741 generic.go:334] "Generic (PLEG): container finished" podID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.947174 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q65b6" event={"ID":"1e20498c-3a09-44e9-a4b1-c2dad8437611","Type":"ContainerDied","Data":"f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.948343 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fkpb5_3c11cbdf-1417-4c8d-b3bb-150f57496287/openstack-network-exporter/0.log" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.948401 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fkpb5" event={"ID":"3c11cbdf-1417-4c8d-b3bb-150f57496287","Type":"ContainerDied","Data":"2049ecd7447da69f0f6a9aaa984099c27e8611d3ad2551ceab940b97a3aae395"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.948423 4741 scope.go:117] "RemoveContainer" containerID="ae21046b33ac269496303f0c89312707199e95ccc69f985f150deaee38b8c665" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.948550 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fkpb5" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.953808 4741 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.953832 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8jc4\" (UniqueName: \"kubernetes.io/projected/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-kube-api-access-f8jc4\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.953841 4741 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fc0eaaf0-97ff-47ea-a15d-0024fbc96439-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.989207 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fkpb5"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.989954 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="af59f98d013533f7b1f00c005e49813bc7d0e7ab704e34a34db4cec3151d2b45" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.989970 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="a979b22c48d4833060b798e59c6172ae0d0fb020aa7081564218be678d051ef8" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.989978 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="878b7f5b4bcf11108cbc1b12fcff16a252c21bb2290b20babdfe2cc60f5f4278" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.989984 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="a68bac92910f13da6b5bd1c4c6fa29f7fec8ba0ea712f3e3c9da7813f47eafee" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.989990 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="5ace69d8ba764865fefb3a17c320882c81b7b6f3fba4329e49c21e37a73ff7f1" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.989996 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="71e20963cd9fad0ee127b4dda01c84dd767efbefb2fcb80296923f8718f1ddbf" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990001 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="49334c9dcc9f0ab359baa344e3fe68ec6f0e5271440eec25dd046555f7cc0e20" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990007 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="8bdeba0cf6a202fe00074a937a1baa94541c1fc21049258f466a2ee65af3a75e" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990013 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="e650c5ffc6ebc6918dc63202d7b9d6aaff80d4b6db1471fd84f3faebfd8b934f" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990027 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="0272c3b7736bc9733f76397bbd76a46beab031e046ac627b3f5b05578332ba13" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990034 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="11c745f1505ad37096a8441c7db5e1f99996efc6719e0f171773f7647d65d6f0" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990040 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="5c99d27cbf271df86190fc67b6bd3129d7c9ebc403240020a83b932781c09aa8" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990047 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="bdcc643cf7b0ba52293b398520b054d5a0af7419de83697c00319b14e82a22f5" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990053 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="0f18bf31b15ac6485b4321250e71d201fadbe48a440243ff5584feadda173ffc" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990115 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"af59f98d013533f7b1f00c005e49813bc7d0e7ab704e34a34db4cec3151d2b45"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990135 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"a979b22c48d4833060b798e59c6172ae0d0fb020aa7081564218be678d051ef8"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990147 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"878b7f5b4bcf11108cbc1b12fcff16a252c21bb2290b20babdfe2cc60f5f4278"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990158 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"a68bac92910f13da6b5bd1c4c6fa29f7fec8ba0ea712f3e3c9da7813f47eafee"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990187 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"5ace69d8ba764865fefb3a17c320882c81b7b6f3fba4329e49c21e37a73ff7f1"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990198 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"71e20963cd9fad0ee127b4dda01c84dd767efbefb2fcb80296923f8718f1ddbf"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990206 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"49334c9dcc9f0ab359baa344e3fe68ec6f0e5271440eec25dd046555f7cc0e20"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990214 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"8bdeba0cf6a202fe00074a937a1baa94541c1fc21049258f466a2ee65af3a75e"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990222 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"e650c5ffc6ebc6918dc63202d7b9d6aaff80d4b6db1471fd84f3faebfd8b934f"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990258 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"0272c3b7736bc9733f76397bbd76a46beab031e046ac627b3f5b05578332ba13"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990268 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"11c745f1505ad37096a8441c7db5e1f99996efc6719e0f171773f7647d65d6f0"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990276 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"5c99d27cbf271df86190fc67b6bd3129d7c9ebc403240020a83b932781c09aa8"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990284 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"bdcc643cf7b0ba52293b398520b054d5a0af7419de83697c00319b14e82a22f5"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.990292 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"0f18bf31b15ac6485b4321250e71d201fadbe48a440243ff5584feadda173ffc"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.993692 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_69f1c2bd-c091-42e8-8810-a04726ce9032/ovsdbserver-sb/0.log" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.993788 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"69f1c2bd-c091-42e8-8810-a04726ce9032","Type":"ContainerDied","Data":"7301a38c44c8ebcd85b5c2b28c98a2a30bc19e665453c9b7f8eeb740f1aa5a22"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.993891 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.994411 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-fkpb5"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:05.995478 4741 scope.go:117] "RemoveContainer" containerID="b10e00d6e3d0b86eee6df15910497c173193cb10140a8e44a1e3e58aa399a2cd" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.009481 4741 generic.go:334] "Generic (PLEG): container finished" podID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerID="5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.009541 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c907b3ee-67ab-4d29-a02d-96137c124ff4","Type":"ContainerDied","Data":"5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.019732 4741 generic.go:334] "Generic (PLEG): container finished" podID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerID="8dccd7363f3ef4f5fc4ea08cd195ce440540c0c08f0954b35a076493bfa65bf4" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.019785 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be00d3f7-3dee-4e64-8559-2a5efe4ceba3","Type":"ContainerDied","Data":"8dccd7363f3ef4f5fc4ea08cd195ce440540c0c08f0954b35a076493bfa65bf4"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.021227 4741 generic.go:334] "Generic (PLEG): container finished" podID="fc0eaaf0-97ff-47ea-a15d-0024fbc96439" containerID="a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead" exitCode=137 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.021320 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.053349 4741 scope.go:117] "RemoveContainer" containerID="814a325d7b6c7b6ac7fd7b9d3bac9dffa54a1007b3dabf568b1a0a21bc488445" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.053692 4741 generic.go:334] "Generic (PLEG): container finished" podID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerID="eac6f805571d33e47abcb8386b1e4a78317268ba124fe10a2e9376e94d5dd506" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.053778 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3b65efd-d659-4b2d-9cee-0c57b08d029f","Type":"ContainerDied","Data":"eac6f805571d33e47abcb8386b1e4a78317268ba124fe10a2e9376e94d5dd506"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.056713 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.062283 4741 generic.go:334] "Generic (PLEG): container finished" podID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerID="f4540013c871a62255548905e7c7a5f8f143d8e4617fe18f7c6f7bdbd529b257" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.062334 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c96956bc-7qhgv" event={"ID":"d3e3a43d-5ff8-491d-916c-3c0b9e31f223","Type":"ContainerDied","Data":"f4540013c871a62255548905e7c7a5f8f143d8e4617fe18f7c6f7bdbd529b257"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.065063 4741 generic.go:334] "Generic (PLEG): container finished" podID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerID="58644df2679ff5827ca80923ddee91fde9b4da2f9eb6e37ae982ca0c9572be55" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.065128 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659f5d9f44-j8zgh" event={"ID":"58cfa8a3-5067-40ee-8e23-803a52d6b1b1","Type":"ContainerDied","Data":"58644df2679ff5827ca80923ddee91fde9b4da2f9eb6e37ae982ca0c9572be55"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.067183 4741 generic.go:334] "Generic (PLEG): container finished" podID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerID="1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.067219 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" event={"ID":"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb","Type":"ContainerDied","Data":"1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.067236 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" event={"ID":"462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb","Type":"ContainerDied","Data":"afe473e033f4be1934d8ed1c53f5d257398a66f8ee46f1776279b0e6cc62bd9c"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.067292 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-55hm7" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.070176 4741 generic.go:334] "Generic (PLEG): container finished" podID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerID="59e07c173f297fc52b58154fee10e2757e48b2c11cd92731bb4e2fc98f386bc7" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.070213 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c9251106-ba81-446b-be7b-51c4ac7f3f81","Type":"ContainerDied","Data":"59e07c173f297fc52b58154fee10e2757e48b2c11cd92731bb4e2fc98f386bc7"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.072555 4741 generic.go:334] "Generic (PLEG): container finished" podID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerID="fc163c8b456c4eae3276be79b996f9a990070e39bb81a710df286eea59af6f36" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.072613 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" event={"ID":"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d","Type":"ContainerDied","Data":"fc163c8b456c4eae3276be79b996f9a990070e39bb81a710df286eea59af6f36"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.081288 4741 generic.go:334] "Generic (PLEG): container finished" podID="b8c0591c-26f3-48eb-9553-e61c932fda9d" containerID="1397bfc7632f5c941d49979998616e28d53946a240b4241566583acbcdd1203e" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.081332 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement1b19-account-delete-shzs7" event={"ID":"b8c0591c-26f3-48eb-9553-e61c932fda9d","Type":"ContainerDied","Data":"1397bfc7632f5c941d49979998616e28d53946a240b4241566583acbcdd1203e"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.081355 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement1b19-account-delete-shzs7" event={"ID":"b8c0591c-26f3-48eb-9553-e61c932fda9d","Type":"ContainerStarted","Data":"f07182d5141bdbfcf1ab901a44d974d5a28fe38c97e289bafec89c29a78ccc77"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.090189 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.097613 4741 generic.go:334] "Generic (PLEG): container finished" podID="2094977c-9460-4dfd-926d-aac495e9bb73" containerID="f74c820b505971e578bc8f81685750786c126b92ec697fda00bad07f4d638490" exitCode=0 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.097688 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron07da-account-delete-vj777" event={"ID":"2094977c-9460-4dfd-926d-aac495e9bb73","Type":"ContainerDied","Data":"f74c820b505971e578bc8f81685750786c126b92ec697fda00bad07f4d638490"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.097709 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron07da-account-delete-vj777" event={"ID":"2094977c-9460-4dfd-926d-aac495e9bb73","Type":"ContainerStarted","Data":"7798d0eace02a25729597b888d6238b301b6b9609096cc84eb27def931df1704"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.113276 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell15b24-account-delete-whtzz" event={"ID":"13785b5b-f629-47ae-8251-331d44b40254","Type":"ContainerStarted","Data":"b620ebe1c97698a6498d508a46e150f1702c9557a9a398daa2eb00b541592b9f"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.124264 4741 generic.go:334] "Generic (PLEG): container finished" podID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerID="e1b73040b554cbd478a51c8b5816e58a0a68a53c90c12ff3ca2f8c2045c814bf" exitCode=143 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.124330 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ade550-0eaf-4518-8696-14daf3f034d4","Type":"ContainerDied","Data":"e1b73040b554cbd478a51c8b5816e58a0a68a53c90c12ff3ca2f8c2045c814bf"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.125998 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6af0-account-delete-8hrbb" event={"ID":"983d2897-24d2-462c-b9f9-427639d2b8f9","Type":"ContainerStarted","Data":"749520f305926a8a1fb56868f7cbd0a33f054e60841c28493f2b2a74e2842488"} Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.288812 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-74d7c6f55-9trnx"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.289308 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-74d7c6f55-9trnx" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-httpd" containerID="cri-o://cce06dd4a20cab8469d6440e5d832029b66e459b52c5ff9b817bb86dbe5b5145" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.289709 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-74d7c6f55-9trnx" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-server" containerID="cri-o://5488046b5c71d7ada89c02641c087812ffa5561b258e4750ab91f6cd07383d28" gracePeriod=30 Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.365202 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" (UID: "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.367971 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.370977 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" (UID: "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.379264 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" (UID: "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.389886 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" (UID: "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.399257 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-config" (OuterVolumeSpecName: "config") pod "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" (UID: "462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.469148 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.469180 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.469191 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.469200 4741 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.738327 4741 scope.go:117] "RemoveContainer" containerID="a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead" Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:06.774786 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:06.774940 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data podName:8e0c02dc-69a9-4e60-b179-0e23842d10a4 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:10.774921572 +0000 UTC m=+1312.422710904 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data") pod "rabbitmq-cell1-server-0" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4") : configmap "rabbitmq-cell1-config-data" not found Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.810118 4741 scope.go:117] "RemoveContainer" containerID="a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead" Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:06.810602 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead\": container with ID starting with a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead not found: ID does not exist" containerID="a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.810644 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead"} err="failed to get container status \"a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead\": rpc error: code = NotFound desc = could not find container \"a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead\": container with ID starting with a61546a832fded68ee4b5d4a6fe92094c05c58a457b72de98c251b9f95dceead not found: ID does not exist" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.810670 4741 scope.go:117] "RemoveContainer" containerID="1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.823169 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement1b19-account-delete-shzs7" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.839365 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.840083 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-55hm7"] Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.843272 4741 scope.go:117] "RemoveContainer" containerID="056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.859058 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-55hm7"] Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:06.881559 4741 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:06.881655 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data podName:df6beb49-03ad-47ef-a9c7-3f37baa6d105 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:10.881640807 +0000 UTC m=+1312.529430139 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data") pod "rabbitmq-server-0" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105") : configmap "rabbitmq-config-data" not found Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.885671 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.903181 4741 scope.go:117] "RemoveContainer" containerID="1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da" Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:06.903923 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da\": container with ID starting with 1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da not found: ID does not exist" containerID="1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.903959 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da"} err="failed to get container status \"1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da\": rpc error: code = NotFound desc = could not find container \"1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da\": container with ID starting with 1663015c0380c82f9dd2c70c953e67805f5532aeae0bf6e7b1d01f8f31f626da not found: ID does not exist" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.903980 4741 scope.go:117] "RemoveContainer" containerID="056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d" Sep 29 19:31:06 crc kubenswrapper[4741]: E0929 19:31:06.908205 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d\": container with ID starting with 056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d not found: ID does not exist" containerID="056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.908254 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d"} err="failed to get container status \"056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d\": rpc error: code = NotFound desc = could not find container \"056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d\": container with ID starting with 056cb369eb2813521e0cfc529da46161c843269834324126feeef2d1bb50bc2d not found: ID does not exist" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.949893 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.982456 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjw59\" (UniqueName: \"kubernetes.io/projected/b8c0591c-26f3-48eb-9553-e61c932fda9d-kube-api-access-wjw59\") pod \"b8c0591c-26f3-48eb-9553-e61c932fda9d\" (UID: \"b8c0591c-26f3-48eb-9553-e61c932fda9d\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.982516 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwdc8\" (UniqueName: \"kubernetes.io/projected/2094977c-9460-4dfd-926d-aac495e9bb73-kube-api-access-xwdc8\") pod \"2094977c-9460-4dfd-926d-aac495e9bb73\" (UID: \"2094977c-9460-4dfd-926d-aac495e9bb73\") " Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.990698 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c0591c-26f3-48eb-9553-e61c932fda9d-kube-api-access-wjw59" (OuterVolumeSpecName: "kube-api-access-wjw59") pod "b8c0591c-26f3-48eb-9553-e61c932fda9d" (UID: "b8c0591c-26f3-48eb-9553-e61c932fda9d"). InnerVolumeSpecName "kube-api-access-wjw59". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:06 crc kubenswrapper[4741]: I0929 19:31:06.995502 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2094977c-9460-4dfd-926d-aac495e9bb73-kube-api-access-xwdc8" (OuterVolumeSpecName: "kube-api-access-xwdc8") pod "2094977c-9460-4dfd-926d-aac495e9bb73" (UID: "2094977c-9460-4dfd-926d-aac495e9bb73"). InnerVolumeSpecName "kube-api-access-xwdc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084333 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhrqj\" (UniqueName: \"kubernetes.io/projected/e3b64f81-4c6d-4d71-8930-109240d8ec42-kube-api-access-rhrqj\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084580 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-config-data\") pod \"f24ec6be-6832-4eee-916c-2237e81bf0f1\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084611 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084642 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-generated\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084665 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-default\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084728 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rvch\" (UniqueName: \"kubernetes.io/projected/f24ec6be-6832-4eee-916c-2237e81bf0f1-kube-api-access-9rvch\") pod \"f24ec6be-6832-4eee-916c-2237e81bf0f1\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084751 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-kolla-config\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084768 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-secrets\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084805 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-combined-ca-bundle\") pod \"f24ec6be-6832-4eee-916c-2237e81bf0f1\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084833 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-galera-tls-certs\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084876 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-operator-scripts\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084913 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-combined-ca-bundle\") pod \"e3b64f81-4c6d-4d71-8930-109240d8ec42\" (UID: \"e3b64f81-4c6d-4d71-8930-109240d8ec42\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084937 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-nova-novncproxy-tls-certs\") pod \"f24ec6be-6832-4eee-916c-2237e81bf0f1\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.084972 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-vencrypt-tls-certs\") pod \"f24ec6be-6832-4eee-916c-2237e81bf0f1\" (UID: \"f24ec6be-6832-4eee-916c-2237e81bf0f1\") " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.085561 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.085588 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwdc8\" (UniqueName: \"kubernetes.io/projected/2094977c-9460-4dfd-926d-aac495e9bb73-kube-api-access-xwdc8\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.085868 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjw59\" (UniqueName: \"kubernetes.io/projected/b8c0591c-26f3-48eb-9553-e61c932fda9d-kube-api-access-wjw59\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.087961 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b64f81-4c6d-4d71-8930-109240d8ec42-kube-api-access-rhrqj" (OuterVolumeSpecName: "kube-api-access-rhrqj") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "kube-api-access-rhrqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.088986 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.089403 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.091565 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.101674 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-secrets" (OuterVolumeSpecName: "secrets") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.107879 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08063773-12db-4af5-bf43-6ff10b1672f7" path="/var/lib/kubelet/pods/08063773-12db-4af5-bf43-6ff10b1672f7/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.108543 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21479272-e739-4a74-a149-8c63a5800975" path="/var/lib/kubelet/pods/21479272-e739-4a74-a149-8c63a5800975/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.109106 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c11cbdf-1417-4c8d-b3bb-150f57496287" path="/var/lib/kubelet/pods/3c11cbdf-1417-4c8d-b3bb-150f57496287/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.109746 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" path="/var/lib/kubelet/pods/462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.110740 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eae36aa-f785-4402-ae16-3d3a68050ebb" path="/var/lib/kubelet/pods/4eae36aa-f785-4402-ae16-3d3a68050ebb/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.111168 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f1730aa-2bc9-4005-88fb-4634745de0aa" path="/var/lib/kubelet/pods/4f1730aa-2bc9-4005-88fb-4634745de0aa/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.111822 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6279162a-35c0-4528-970c-f1130fda4fcc" path="/var/lib/kubelet/pods/6279162a-35c0-4528-970c-f1130fda4fcc/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.112963 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="634aed84-d361-421a-a490-fc9db19afe27" path="/var/lib/kubelet/pods/634aed84-d361-421a-a490-fc9db19afe27/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.113576 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" path="/var/lib/kubelet/pods/69f1c2bd-c091-42e8-8810-a04726ce9032/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.114142 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f495781-1870-461e-b12d-ddf347b0a73c" path="/var/lib/kubelet/pods/8f495781-1870-461e-b12d-ddf347b0a73c/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.115028 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91807a37-17a2-451a-bc6a-69f7cd09707c" path="/var/lib/kubelet/pods/91807a37-17a2-451a-bc6a-69f7cd09707c/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.115524 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3f71de-d7bd-4c9c-b44c-63016e612edc" path="/var/lib/kubelet/pods/9a3f71de-d7bd-4c9c-b44c-63016e612edc/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.116724 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" path="/var/lib/kubelet/pods/f5597514-2fd8-4d92-b115-05b66894ea94/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.117299 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f830c8bb-5f61-4a1f-bb46-d59a64d110f9" path="/var/lib/kubelet/pods/f830c8bb-5f61-4a1f-bb46-d59a64d110f9/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.119505 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8909956-17f7-4fa4-805b-e36c4dd756a5" path="/var/lib/kubelet/pods/f8909956-17f7-4fa4-805b-e36c4dd756a5/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.120233 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9e005f0-c926-4dee-a78a-2d5da494bd93" path="/var/lib/kubelet/pods/f9e005f0-c926-4dee-a78a-2d5da494bd93/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.120997 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0eaaf0-97ff-47ea-a15d-0024fbc96439" path="/var/lib/kubelet/pods/fc0eaaf0-97ff-47ea-a15d-0024fbc96439/volumes" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.145174 4741 generic.go:334] "Generic (PLEG): container finished" podID="13785b5b-f629-47ae-8251-331d44b40254" containerID="b14eefa9c9c65e040f41485625e023675c399fa3746763f7c230f6e4192d6b5e" exitCode=1 Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.145732 4741 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell15b24-account-delete-whtzz" secret="" err="secret \"galera-openstack-cell1-dockercfg-nkn5x\" not found" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.145761 4741 scope.go:117] "RemoveContainer" containerID="b14eefa9c9c65e040f41485625e023675c399fa3746763f7c230f6e4192d6b5e" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.172778 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "mysql-db") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.182515 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f24ec6be-6832-4eee-916c-2237e81bf0f1-kube-api-access-9rvch" (OuterVolumeSpecName: "kube-api-access-9rvch") pod "f24ec6be-6832-4eee-916c-2237e81bf0f1" (UID: "f24ec6be-6832-4eee-916c-2237e81bf0f1"). InnerVolumeSpecName "kube-api-access-9rvch". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189087 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhrqj\" (UniqueName: \"kubernetes.io/projected/e3b64f81-4c6d-4d71-8930-109240d8ec42-kube-api-access-rhrqj\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189133 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189143 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-generated\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189151 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-config-data-default\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189160 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rvch\" (UniqueName: \"kubernetes.io/projected/f24ec6be-6832-4eee-916c-2237e81bf0f1-kube-api-access-9rvch\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189169 4741 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-kolla-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189176 4741 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.189198 4741 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b64f81-4c6d-4d71-8930-109240d8ec42-operator-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.196333 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f24ec6be-6832-4eee-916c-2237e81bf0f1" (UID: "f24ec6be-6832-4eee-916c-2237e81bf0f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.214634 4741 generic.go:334] "Generic (PLEG): container finished" podID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerID="15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb" exitCode=0 Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.214795 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.225979 4741 generic.go:334] "Generic (PLEG): container finished" podID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerID="5488046b5c71d7ada89c02641c087812ffa5561b258e4750ab91f6cd07383d28" exitCode=0 Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.226021 4741 generic.go:334] "Generic (PLEG): container finished" podID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerID="cce06dd4a20cab8469d6440e5d832029b66e459b52c5ff9b817bb86dbe5b5145" exitCode=0 Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.227687 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement1b19-account-delete-shzs7" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.239988 4741 generic.go:334] "Generic (PLEG): container finished" podID="9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62" containerID="aca50d7d0a09ea3de4b166aafd5c60114abc7d8e450c9b107c4c352810e4b20a" exitCode=0 Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.271137 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "f24ec6be-6832-4eee-916c-2237e81bf0f1" (UID: "f24ec6be-6832-4eee-916c-2237e81bf0f1"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.299844 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.299866 4741 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.301459 4741 generic.go:334] "Generic (PLEG): container finished" podID="f24ec6be-6832-4eee-916c-2237e81bf0f1" containerID="e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3" exitCode=0 Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.301635 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.306858 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.310203 4741 generic.go:334] "Generic (PLEG): container finished" podID="983d2897-24d2-462c-b9f9-427639d2b8f9" containerID="d48e3684ebfff17addc725043e0493445f56a413b0de369662ab33c743879d3d" exitCode=0 Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.333047 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.352335 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.404488 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.404718 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.409318 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "f24ec6be-6832-4eee-916c-2237e81bf0f1" (UID: "f24ec6be-6832-4eee-916c-2237e81bf0f1"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.412978 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-config-data" (OuterVolumeSpecName: "config-data") pod "f24ec6be-6832-4eee-916c-2237e81bf0f1" (UID: "f24ec6be-6832-4eee-916c-2237e81bf0f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.441271 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "e3b64f81-4c6d-4d71-8930-109240d8ec42" (UID: "e3b64f81-4c6d-4d71-8930-109240d8ec42"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.506011 4741 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b64f81-4c6d-4d71-8930-109240d8ec42-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.506040 4741 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: I0929 19:31:07.506050 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ec6be-6832-4eee-916c-2237e81bf0f1-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.522289 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.551376 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.568539 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.568625 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="ovn-northd" Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.762130 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.781045 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.827535 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.827620 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="93c5456d-7a59-4991-ad3d-58b04db78b24" containerName="nova-cell1-conductor-conductor" Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.931986 4741 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.932250 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:11.932237572 +0000 UTC m=+1313.580026904 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-config" not found Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.933094 4741 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Sep 29 19:31:07 crc kubenswrapper[4741]: E0929 19:31:07.933119 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:11.933111859 +0000 UTC m=+1313.580901191 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-httpd-config" not found Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.326665 4741 generic.go:334] "Generic (PLEG): container finished" podID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerID="4b30dc6f2664d0596e57ddb9855aca996f10f7caab7b0df97ee564ddcc5bc232" exitCode=0 Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.330153 4741 generic.go:334] "Generic (PLEG): container finished" podID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerID="5d7c9dcdb6e6631c524f68a99886213580f717eb067525088f01ff7e8451b3e4" exitCode=0 Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.335893 4741 generic.go:334] "Generic (PLEG): container finished" podID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerID="b426eb53e2825c670095e43996bf58ecec4760a4bafd71a4d7304962625a63c7" exitCode=0 Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.342203 4741 generic.go:334] "Generic (PLEG): container finished" podID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerID="c9909a6976e00256c9ff763bed1a932c9df0fd365ff4dabc64990d63e5be6cf4" exitCode=0 Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.343533 4741 generic.go:334] "Generic (PLEG): container finished" podID="93c5456d-7a59-4991-ad3d-58b04db78b24" containerID="aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070" exitCode=0 Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.345956 4741 generic.go:334] "Generic (PLEG): container finished" podID="2be2676f-d882-424e-8a49-d8bffc23773e" containerID="88822bc2858601e3da0fba4f6a6ca50dcd1d5c1fe288573a193d9458b638d155" exitCode=0 Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.348468 4741 generic.go:334] "Generic (PLEG): container finished" podID="9d18b4c1-d423-4840-97ff-c322272c3aa3" containerID="e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb" exitCode=0 Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.434029 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb is running failed: container process not found" containerID="e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.434944 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb is running failed: container process not found" containerID="e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.435197 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb is running failed: container process not found" containerID="e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.435224 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9d18b4c1-d423-4840-97ff-c322272c3aa3" containerName="nova-scheduler-scheduler" Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.470302 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.162:8776/healthcheck\": read tcp 10.217.0.2:37906->10.217.0.162:8776: read: connection reset by peer" Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.510340 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:60760->10.217.0.200:8775: read: connection reset by peer" Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.510724 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:60748->10.217.0.200:8775: read: connection reset by peer" Sep 29 19:31:08 crc kubenswrapper[4741]: I0929 19:31:08.922264 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-68snt" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerName="ovn-controller" probeResult="failure" output=< Sep 29 19:31:08 crc kubenswrapper[4741]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Sep 29 19:31:08 crc kubenswrapper[4741]: > Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.931190 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.931546 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.932456 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.932709 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.932747 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.934070 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.936663 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:08 crc kubenswrapper[4741]: E0929 19:31:08.936691 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.004988 4741 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.919s" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005026 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell15b24-account-delete-whtzz" event={"ID":"13785b5b-f629-47ae-8251-331d44b40254","Type":"ContainerDied","Data":"b14eefa9c9c65e040f41485625e023675c399fa3746763f7c230f6e4192d6b5e"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005125 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005143 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005153 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005165 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e3b64f81-4c6d-4d71-8930-109240d8ec42","Type":"ContainerDied","Data":"15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005181 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-7fjg8"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005193 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-7fjg8"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005205 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-s9tpc"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005214 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e3b64f81-4c6d-4d71-8930-109240d8ec42","Type":"ContainerDied","Data":"fabcaa4ef29605769073643955da8e00621af8d671f6bf30dc21156315f6dec8"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005237 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d7c6f55-9trnx" event={"ID":"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe","Type":"ContainerDied","Data":"5488046b5c71d7ada89c02641c087812ffa5561b258e4750ab91f6cd07383d28"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005252 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d7c6f55-9trnx" event={"ID":"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe","Type":"ContainerDied","Data":"cce06dd4a20cab8469d6440e5d832029b66e459b52c5ff9b817bb86dbe5b5145"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005262 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement1b19-account-delete-shzs7" event={"ID":"b8c0591c-26f3-48eb-9553-e61c932fda9d","Type":"ContainerDied","Data":"f07182d5141bdbfcf1ab901a44d974d5a28fe38c97e289bafec89c29a78ccc77"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005272 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi7d7d-account-delete-wklvh" event={"ID":"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62","Type":"ContainerDied","Data":"aca50d7d0a09ea3de4b166aafd5c60114abc7d8e450c9b107c4c352810e4b20a"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005282 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi7d7d-account-delete-wklvh" event={"ID":"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62","Type":"ContainerStarted","Data":"5485f4dde0e8c2aada6c318857da0cb683be25b7b84cf9b037cc9e5d71fb2fbf"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005293 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-s9tpc"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005304 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f24ec6be-6832-4eee-916c-2237e81bf0f1","Type":"ContainerDied","Data":"e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005317 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f24ec6be-6832-4eee-916c-2237e81bf0f1","Type":"ContainerDied","Data":"78582c14d813d4691bde7fb531ce55242e16355d419af06008602af4258d247a"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005332 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone81dc-account-delete-blpg9"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005460 4741 scope.go:117] "RemoveContainer" containerID="15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005531 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-central-agent" containerID="cri-o://64aaaacdbcb9f52035038a67edb1624996ab0fb313a96a375112066aa4d0c3a6" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005693 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="proxy-httpd" containerID="cri-o://167d2c7f919eb9ab445a2017a6763de4f7c20dd26fb277979f6106c71a0731b5" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005762 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="sg-core" containerID="cri-o://113041bee656cb97e263bb0fbf9286ccb77bdf43fa28dcf029dc790fd12a60d8" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.005801 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-notification-agent" containerID="cri-o://3d41a9cc13ba07eab14e449222ca1eba4bc2f913c9d707b7f2f9fe2953779503" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.007074 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" containerName="kube-state-metrics" containerID="cri-o://7720681edbf8321fa29bc81ceddd38f098ef1fa29db00e81fd9806812c014bd7" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.007133 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" containerName="memcached" containerID="cri-o://7c71b4d390feacc6365d3ebfe09f6e223b707ee503a77ad41f8f50b865956428" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.008340 4741 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/keystone-677b8f6f88-c7kf7" secret="" err="secret \"keystone-keystone-dockercfg-kg6nl\" not found" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.009783 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c0591c-26f3-48eb-9553-e61c932fda9d" containerName="mariadb-account-delete" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.009829 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c0591c-26f3-48eb-9553-e61c932fda9d" containerName="mariadb-account-delete" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.009863 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2094977c-9460-4dfd-926d-aac495e9bb73" containerName="mariadb-account-delete" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.009870 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2094977c-9460-4dfd-926d-aac495e9bb73" containerName="mariadb-account-delete" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.009901 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="ovsdbserver-nb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.009908 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="ovsdbserver-nb" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.009923 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c11cbdf-1417-4c8d-b3bb-150f57496287" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.009930 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c11cbdf-1417-4c8d-b3bb-150f57496287" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.009953 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerName="init" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.009960 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerName="init" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.009972 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.009979 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.009999 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="ovsdbserver-sb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010005 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="ovsdbserver-sb" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.010017 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f24ec6be-6832-4eee-916c-2237e81bf0f1" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010024 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f24ec6be-6832-4eee-916c-2237e81bf0f1" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.010034 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerName="dnsmasq-dns" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010040 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerName="dnsmasq-dns" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.010050 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerName="galera" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010057 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerName="galera" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.010067 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerName="mysql-bootstrap" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010076 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerName="mysql-bootstrap" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.010090 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010096 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010335 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010343 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b64f81-4c6d-4d71-8930-109240d8ec42" containerName="galera" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010356 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c0591c-26f3-48eb-9553-e61c932fda9d" containerName="mariadb-account-delete" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010366 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f1c2bd-c091-42e8-8810-a04726ce9032" containerName="ovsdbserver-sb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010372 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2094977c-9460-4dfd-926d-aac495e9bb73" containerName="mariadb-account-delete" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010401 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c11cbdf-1417-4c8d-b3bb-150f57496287" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010410 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f24ec6be-6832-4eee-916c-2237e81bf0f1" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010420 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="ovsdbserver-nb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010450 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="462c2df4-bb0b-471a-8f1c-a0e15a7fe1eb" containerName="dnsmasq-dns" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010459 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5597514-2fd8-4d92-b115-05b66894ea94" containerName="openstack-network-exporter" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.010990 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron07da-account-delete-vj777" event={"ID":"2094977c-9460-4dfd-926d-aac495e9bb73","Type":"ContainerDied","Data":"7798d0eace02a25729597b888d6238b301b6b9609096cc84eb27def931df1704"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011018 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-677b8f6f88-c7kf7"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011035 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6af0-account-delete-8hrbb" event={"ID":"983d2897-24d2-462c-b9f9-427639d2b8f9","Type":"ContainerDied","Data":"d48e3684ebfff17addc725043e0493445f56a413b0de369662ab33c743879d3d"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011050 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone81dc-account-delete-blpg9"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011062 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" event={"ID":"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d","Type":"ContainerDied","Data":"4b30dc6f2664d0596e57ddb9855aca996f10f7caab7b0df97ee564ddcc5bc232"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011073 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" event={"ID":"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d","Type":"ContainerDied","Data":"b42500dea64ea4de76dc6f41aa6240da449020927e24ac47aca39f94f9c04e59"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011082 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b42500dea64ea4de76dc6f41aa6240da449020927e24ac47aca39f94f9c04e59" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011090 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c96956bc-7qhgv" event={"ID":"d3e3a43d-5ff8-491d-916c-3c0b9e31f223","Type":"ContainerDied","Data":"5d7c9dcdb6e6631c524f68a99886213580f717eb067525088f01ff7e8451b3e4"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011101 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011112 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-hqmd7"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011122 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-hqmd7"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011133 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6af0-account-delete-8hrbb" event={"ID":"983d2897-24d2-462c-b9f9-427639d2b8f9","Type":"ContainerDied","Data":"749520f305926a8a1fb56868f7cbd0a33f054e60841c28493f2b2a74e2842488"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011141 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="749520f305926a8a1fb56868f7cbd0a33f054e60841c28493f2b2a74e2842488" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011149 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-81dc-account-create-ghmk4"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011157 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c9251106-ba81-446b-be7b-51c4ac7f3f81","Type":"ContainerDied","Data":"b426eb53e2825c670095e43996bf58ecec4760a4bafd71a4d7304962625a63c7"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011167 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-81dc-account-create-ghmk4"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011177 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d7c6f55-9trnx" event={"ID":"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe","Type":"ContainerDied","Data":"7c8e4008fd8e4be8cdfa645117d0791e6517e3b394e16325326266514ea489ee"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011187 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c8e4008fd8e4be8cdfa645117d0791e6517e3b394e16325326266514ea489ee" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011194 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ade550-0eaf-4518-8696-14daf3f034d4","Type":"ContainerDied","Data":"c9909a6976e00256c9ff763bed1a932c9df0fd365ff4dabc64990d63e5be6cf4"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011203 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93c5456d-7a59-4991-ad3d-58b04db78b24","Type":"ContainerDied","Data":"aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011214 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone81dc-account-delete-blpg9"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011225 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi7d7d-account-delete-wklvh" event={"ID":"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62","Type":"ContainerDied","Data":"5485f4dde0e8c2aada6c318857da0cb683be25b7b84cf9b037cc9e5d71fb2fbf"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011233 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5485f4dde0e8c2aada6c318857da0cb683be25b7b84cf9b037cc9e5d71fb2fbf" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011241 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-gcx8j"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011250 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c686d89f8-xwpjv" event={"ID":"2be2676f-d882-424e-8a49-d8bffc23773e","Type":"ContainerDied","Data":"88822bc2858601e3da0fba4f6a6ca50dcd1d5c1fe288573a193d9458b638d155"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011265 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d18b4c1-d423-4840-97ff-c322272c3aa3","Type":"ContainerDied","Data":"e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011287 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-gcx8j"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011298 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell15b24-account-delete-whtzz"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011309 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5b24-account-create-bhddd"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.011318 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5b24-account-create-bhddd"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.013604 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone81dc-account-delete-blpg9" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.023045 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.059988 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q8fl\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-kube-api-access-8q8fl\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.060083 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-public-tls-certs\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.060135 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-combined-ca-bundle\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.060192 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-internal-tls-certs\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.060255 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-log-httpd\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.060338 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-config-data\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.060368 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-etc-swift\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.060408 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-run-httpd\") pod \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\" (UID: \"2efb9fff-2e44-4257-bbe6-e5dccedcdcbe\") " Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.061482 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.061531 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:09.5615174 +0000 UTC m=+1311.209306722 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.063835 4741 secret.go:188] Couldn't get secret openstack/keystone-config-data: secret "keystone-config-data" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.063925 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:09.563893335 +0000 UTC m=+1311.211682667 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-config-data" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.063965 4741 secret.go:188] Couldn't get secret openstack/keystone-scripts: secret "keystone-scripts" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.063983 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:09.563977477 +0000 UTC m=+1311.211766809 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-scripts" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.066809 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.066844 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:09.566835467 +0000 UTC m=+1311.214624799 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.072074 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.076819 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.107933 4741 scope.go:117] "RemoveContainer" containerID="a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.108592 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a8d0551-6f96-4651-908d-3a048f691ddb" path="/var/lib/kubelet/pods/2a8d0551-6f96-4651-908d-3a048f691ddb/volumes" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.109251 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d0b8a94-1812-451f-9800-c46d88559954" path="/var/lib/kubelet/pods/4d0b8a94-1812-451f-9800-c46d88559954/volumes" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.109964 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e4efdb3-fa44-4d1b-b62c-007252565734" path="/var/lib/kubelet/pods/9e4efdb3-fa44-4d1b-b62c-007252565734/volumes" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.110880 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce0c5a54-c02f-4f97-a0ea-54383f8b792c" path="/var/lib/kubelet/pods/ce0c5a54-c02f-4f97-a0ea-54383f8b792c/volumes" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.112647 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1cf524c-6534-40d5-83eb-5a0b75684622" path="/var/lib/kubelet/pods/d1cf524c-6534-40d5-83eb-5a0b75684622/volumes" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.113915 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d38d17bd-4e0b-489d-8b4a-2047a1004e25" path="/var/lib/kubelet/pods/d38d17bd-4e0b-489d-8b4a-2047a1004e25/volumes" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.116183 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-kube-api-access-8q8fl" (OuterVolumeSpecName: "kube-api-access-8q8fl") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "kube-api-access-8q8fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.123476 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.163054 4741 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.163082 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.163091 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q8fl\" (UniqueName: \"kubernetes.io/projected/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-kube-api-access-8q8fl\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.163100 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.164459 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi7d7d-account-delete-wklvh" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.179737 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance6af0-account-delete-8hrbb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.201272 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.204253 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement1b19-account-delete-shzs7"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.217260 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.257839 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement1b19-account-delete-shzs7"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.262651 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.265756 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnhdk\" (UniqueName: \"kubernetes.io/projected/9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62-kube-api-access-xnhdk\") pod \"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62\" (UID: \"9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.275586 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62-kube-api-access-xnhdk" (OuterVolumeSpecName: "kube-api-access-xnhdk") pod "9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62" (UID: "9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62"). InnerVolumeSpecName "kube-api-access-xnhdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.278258 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerName="galera" containerID="cri-o://535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.280943 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-config-data" (OuterVolumeSpecName: "config-data") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.283873 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.305098 4741 scope.go:117] "RemoveContainer" containerID="15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.306533 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb\": container with ID starting with 15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb not found: ID does not exist" containerID="15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.306582 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb"} err="failed to get container status \"15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb\": rpc error: code = NotFound desc = could not find container \"15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb\": container with ID starting with 15bfcb97e06247641b3bbed1dc19f6f25d1d7cb6fefeb0b22c8d230f308a6ceb not found: ID does not exist" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.306611 4741 scope.go:117] "RemoveContainer" containerID="a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.307430 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c\": container with ID starting with a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c not found: ID does not exist" containerID="a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.307474 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c"} err="failed to get container status \"a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c\": rpc error: code = NotFound desc = could not find container \"a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c\": container with ID starting with a5f58d4088205fd0f4452e45e84de8750b98c8452219288ca4d3c01139579f0c not found: ID does not exist" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.307506 4741 scope.go:117] "RemoveContainer" containerID="1397bfc7632f5c941d49979998616e28d53946a240b4241566583acbcdd1203e" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.316254 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.350667 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.359322 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.373030 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-564rb\" (UniqueName: \"kubernetes.io/projected/983d2897-24d2-462c-b9f9-427639d2b8f9-kube-api-access-564rb\") pod \"983d2897-24d2-462c-b9f9-427639d2b8f9\" (UID: \"983d2897-24d2-462c-b9f9-427639d2b8f9\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.373120 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data\") pod \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.373171 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-combined-ca-bundle\") pod \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.373463 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9txfg\" (UniqueName: \"kubernetes.io/projected/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-kube-api-access-9txfg\") pod \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.373499 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q8lc\" (UniqueName: \"kubernetes.io/projected/9d18b4c1-d423-4840-97ff-c322272c3aa3-kube-api-access-2q8lc\") pod \"9d18b4c1-d423-4840-97ff-c322272c3aa3\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.373810 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-combined-ca-bundle\") pod \"9d18b4c1-d423-4840-97ff-c322272c3aa3\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.375535 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-logs\") pod \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.376421 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-combined-ca-bundle\") pod \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.376830 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data\") pod \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.377200 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data-custom\") pod \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\" (UID: \"c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.377637 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-925md\" (UniqueName: \"kubernetes.io/projected/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-kube-api-access-925md\") pod \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.378497 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-logs\") pod \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.378659 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-config-data\") pod \"9d18b4c1-d423-4840-97ff-c322272c3aa3\" (UID: \"9d18b4c1-d423-4840-97ff-c322272c3aa3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.378764 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data-custom\") pod \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\" (UID: \"d3e3a43d-5ff8-491d-916c-3c0b9e31f223\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.376284 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-logs" (OuterVolumeSpecName: "logs") pod "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" (UID: "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.380242 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/983d2897-24d2-462c-b9f9-427639d2b8f9-kube-api-access-564rb" (OuterVolumeSpecName: "kube-api-access-564rb") pod "983d2897-24d2-462c-b9f9-427639d2b8f9" (UID: "983d2897-24d2-462c-b9f9-427639d2b8f9"). InnerVolumeSpecName "kube-api-access-564rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.380558 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-logs" (OuterVolumeSpecName: "logs") pod "d3e3a43d-5ff8-491d-916c-3c0b9e31f223" (UID: "d3e3a43d-5ff8-491d-916c-3c0b9e31f223"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.376713 4741 generic.go:334] "Generic (PLEG): container finished" podID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerID="5747543db2d4686ff0ce2d66484ba05733793d16bc7f1a3c6bf312799d425845" exitCode=0 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386250 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnhdk\" (UniqueName: \"kubernetes.io/projected/9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62-kube-api-access-xnhdk\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386347 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-564rb\" (UniqueName: \"kubernetes.io/projected/983d2897-24d2-462c-b9f9-427639d2b8f9-kube-api-access-564rb\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386435 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386513 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386575 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386639 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386759 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.376740 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3b65efd-d659-4b2d-9cee-0c57b08d029f","Type":"ContainerDied","Data":"5747543db2d4686ff0ce2d66484ba05733793d16bc7f1a3c6bf312799d425845"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.386925 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3b65efd-d659-4b2d-9cee-0c57b08d029f","Type":"ContainerDied","Data":"6999f3d4635471e5c45211f770e66bf963e5812a4d3ce22e0299954d1838c487"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.387461 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6999f3d4635471e5c45211f770e66bf963e5812a4d3ce22e0299954d1838c487" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.387557 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.388962 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d3e3a43d-5ff8-491d-916c-3c0b9e31f223" (UID: "d3e3a43d-5ff8-491d-916c-3c0b9e31f223"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.394969 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c686d89f8-xwpjv" event={"ID":"2be2676f-d882-424e-8a49-d8bffc23773e","Type":"ContainerDied","Data":"fe9ff92122e26d5f80efee609332fb9b0ee005e5abfc50bf9d3618a7103d1922"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.395026 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe9ff92122e26d5f80efee609332fb9b0ee005e5abfc50bf9d3618a7103d1922" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.396758 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d18b4c1-d423-4840-97ff-c322272c3aa3-kube-api-access-2q8lc" (OuterVolumeSpecName: "kube-api-access-2q8lc") pod "9d18b4c1-d423-4840-97ff-c322272c3aa3" (UID: "9d18b4c1-d423-4840-97ff-c322272c3aa3"). InnerVolumeSpecName "kube-api-access-2q8lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.399200 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.400210 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d18b4c1-d423-4840-97ff-c322272c3aa3" (UID: "9d18b4c1-d423-4840-97ff-c322272c3aa3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.400832 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" (UID: "2efb9fff-2e44-4257-bbe6-e5dccedcdcbe"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.403926 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-kube-api-access-925md" (OuterVolumeSpecName: "kube-api-access-925md") pod "d3e3a43d-5ff8-491d-916c-3c0b9e31f223" (UID: "d3e3a43d-5ff8-491d-916c-3c0b9e31f223"). InnerVolumeSpecName "kube-api-access-925md". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.410771 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" (UID: "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.414688 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" (UID: "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.416662 4741 generic.go:334] "Generic (PLEG): container finished" podID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerID="167d2c7f919eb9ab445a2017a6763de4f7c20dd26fb277979f6106c71a0731b5" exitCode=0 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.416739 4741 generic.go:334] "Generic (PLEG): container finished" podID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerID="113041bee656cb97e263bb0fbf9286ccb77bdf43fa28dcf029dc790fd12a60d8" exitCode=2 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.416822 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerDied","Data":"167d2c7f919eb9ab445a2017a6763de4f7c20dd26fb277979f6106c71a0731b5"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.416941 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerDied","Data":"113041bee656cb97e263bb0fbf9286ccb77bdf43fa28dcf029dc790fd12a60d8"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.422586 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-kube-api-access-9txfg" (OuterVolumeSpecName: "kube-api-access-9txfg") pod "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" (UID: "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d"). InnerVolumeSpecName "kube-api-access-9txfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.428521 4741 generic.go:334] "Generic (PLEG): container finished" podID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerID="040aa5ec5b498aa5b60ba516e013b9839f5b611e8d7d5311e0be82ba0ff8b3c7" exitCode=0 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.428618 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be00d3f7-3dee-4e64-8559-2a5efe4ceba3","Type":"ContainerDied","Data":"040aa5ec5b498aa5b60ba516e013b9839f5b611e8d7d5311e0be82ba0ff8b3c7"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.428647 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be00d3f7-3dee-4e64-8559-2a5efe4ceba3","Type":"ContainerDied","Data":"d0f7b1610351d19418e7412967a6095f0b940278fd9aaf2840335b3e894dc531"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.428658 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0f7b1610351d19418e7412967a6095f0b940278fd9aaf2840335b3e894dc531" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.432564 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d18b4c1-d423-4840-97ff-c322272c3aa3","Type":"ContainerDied","Data":"893ae0f43d22e06b970470419b3519a6562eae482a28106ea379115dd697272d"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.432650 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.435940 4741 generic.go:334] "Generic (PLEG): container finished" podID="5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" containerID="7720681edbf8321fa29bc81ceddd38f098ef1fa29db00e81fd9806812c014bd7" exitCode=2 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.435985 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a","Type":"ContainerDied","Data":"7720681edbf8321fa29bc81ceddd38f098ef1fa29db00e81fd9806812c014bd7"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.447369 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ade550-0eaf-4518-8696-14daf3f034d4","Type":"ContainerDied","Data":"bcbc7131e1c30527aba2565c71c4ad302351c56425a915b89798715ebb84417f"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.447484 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcbc7131e1c30527aba2565c71c4ad302351c56425a915b89798715ebb84417f" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.448287 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3e3a43d-5ff8-491d-916c-3c0b9e31f223" (UID: "d3e3a43d-5ff8-491d-916c-3c0b9e31f223"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.450028 4741 generic.go:334] "Generic (PLEG): container finished" podID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerID="dfd2422ee682a56f33a264c928c2202071429b72837518805278e82828695484" exitCode=0 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.450084 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659f5d9f44-j8zgh" event={"ID":"58cfa8a3-5067-40ee-8e23-803a52d6b1b1","Type":"ContainerDied","Data":"dfd2422ee682a56f33a264c928c2202071429b72837518805278e82828695484"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.450112 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659f5d9f44-j8zgh" event={"ID":"58cfa8a3-5067-40ee-8e23-803a52d6b1b1","Type":"ContainerDied","Data":"dab1bbdc6afa761fe9a849e8ae5ff2eb85c04e27d9ed56440a407f6e2e4c7bf5"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.450143 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dab1bbdc6afa761fe9a849e8ae5ff2eb85c04e27d9ed56440a407f6e2e4c7bf5" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.450079 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data" (OuterVolumeSpecName: "config-data") pod "d3e3a43d-5ff8-491d-916c-3c0b9e31f223" (UID: "d3e3a43d-5ff8-491d-916c-3c0b9e31f223"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.458182 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-config-data" (OuterVolumeSpecName: "config-data") pod "9d18b4c1-d423-4840-97ff-c322272c3aa3" (UID: "9d18b4c1-d423-4840-97ff-c322272c3aa3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.466638 4741 generic.go:334] "Generic (PLEG): container finished" podID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerID="e3ba9d353708ea132687e9ee72cfd9472209753189bc1863b2bf80c6d632fdd9" exitCode=0 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.466688 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b0a22c3-d524-422d-8692-14d7a16a418f","Type":"ContainerDied","Data":"e3ba9d353708ea132687e9ee72cfd9472209753189bc1863b2bf80c6d632fdd9"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.466709 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b0a22c3-d524-422d-8692-14d7a16a418f","Type":"ContainerDied","Data":"376e5a13a7a12be380fb1f9484cf4feabbedc5762e3ae2e1369aadd1393f5203"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.466864 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="376e5a13a7a12be380fb1f9484cf4feabbedc5762e3ae2e1369aadd1393f5203" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.468739 4741 generic.go:334] "Generic (PLEG): container finished" podID="13785b5b-f629-47ae-8251-331d44b40254" containerID="fe1574bb5521288388babc106992a52d9ec3cca29d6d836f9f0d16b9960f0890" exitCode=1 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.468774 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell15b24-account-delete-whtzz" event={"ID":"13785b5b-f629-47ae-8251-331d44b40254","Type":"ContainerDied","Data":"fe1574bb5521288388babc106992a52d9ec3cca29d6d836f9f0d16b9960f0890"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.472256 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c9251106-ba81-446b-be7b-51c4ac7f3f81","Type":"ContainerDied","Data":"ef2b713f3984fe037901c5d1baa89b3295575d3b2815af686cbb1fa59e043a63"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.472286 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef2b713f3984fe037901c5d1baa89b3295575d3b2815af686cbb1fa59e043a63" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.474539 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93c5456d-7a59-4991-ad3d-58b04db78b24","Type":"ContainerDied","Data":"4c1e39c6e99eb3b23855e0ffbebc623383253f90eb1ac4a3ce9ccec667248510"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.474599 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c1e39c6e99eb3b23855e0ffbebc623383253f90eb1ac4a3ce9ccec667248510" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.477544 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-677b8f6f88-c7kf7" podUID="c81db200-8fff-4a29-952c-4f8609dc8ac8" containerName="keystone-api" containerID="cri-o://18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34" gracePeriod=30 Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.477596 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi7d7d-account-delete-wklvh" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.477651 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8567fd5894-dwdtp" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.478237 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c96956bc-7qhgv" event={"ID":"d3e3a43d-5ff8-491d-916c-3c0b9e31f223","Type":"ContainerDied","Data":"f524324ad5b249dd4c548da49178acfd62c8196cee5665981e9ad768742d315b"} Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.478517 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c96956bc-7qhgv" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.480293 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74d7c6f55-9trnx" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.482940 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance6af0-account-delete-8hrbb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.483312 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data" (OuterVolumeSpecName: "config-data") pod "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" (UID: "c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.490822 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9txfg\" (UniqueName: \"kubernetes.io/projected/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-kube-api-access-9txfg\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.492017 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q8lc\" (UniqueName: \"kubernetes.io/projected/9d18b4c1-d423-4840-97ff-c322272c3aa3-kube-api-access-2q8lc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.492040 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498531 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498550 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498575 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498587 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498598 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-925md\" (UniqueName: \"kubernetes.io/projected/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-kube-api-access-925md\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498607 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d18b4c1-d423-4840-97ff-c322272c3aa3-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498615 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3e3a43d-5ff8-491d-916c-3c0b9e31f223-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498624 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.498633 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.600965 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.601194 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:10.601176719 +0000 UTC m=+1312.248966051 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.601602 4741 secret.go:188] Couldn't get secret openstack/keystone-scripts: secret "keystone-scripts" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.601069 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.601106 4741 secret.go:188] Couldn't get secret openstack/keystone-config-data: secret "keystone-config-data" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.602784 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:10.602762809 +0000 UTC m=+1312.250552191 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-scripts" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.602812 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:10.6028057 +0000 UTC m=+1312.250595032 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.602824 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:10.602819431 +0000 UTC m=+1312.250608763 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-config-data" not found Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.646598 4741 scope.go:117] "RemoveContainer" containerID="e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.651917 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.703856 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f8p4\" (UniqueName: \"kubernetes.io/projected/93c5456d-7a59-4991-ad3d-58b04db78b24-kube-api-access-6f8p4\") pod \"93c5456d-7a59-4991-ad3d-58b04db78b24\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.704001 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-combined-ca-bundle\") pod \"93c5456d-7a59-4991-ad3d-58b04db78b24\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.704054 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-config-data\") pod \"93c5456d-7a59-4991-ad3d-58b04db78b24\" (UID: \"93c5456d-7a59-4991-ad3d-58b04db78b24\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.709667 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c5456d-7a59-4991-ad3d-58b04db78b24-kube-api-access-6f8p4" (OuterVolumeSpecName: "kube-api-access-6f8p4") pod "93c5456d-7a59-4991-ad3d-58b04db78b24" (UID: "93c5456d-7a59-4991-ad3d-58b04db78b24"). InnerVolumeSpecName "kube-api-access-6f8p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.710274 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi7d7d-account-delete-wklvh"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.711471 4741 scope.go:117] "RemoveContainer" containerID="e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.712674 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:31:09 crc kubenswrapper[4741]: E0929 19:31:09.712764 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3\": container with ID starting with e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3 not found: ID does not exist" containerID="e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.712791 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3"} err="failed to get container status \"e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3\": rpc error: code = NotFound desc = could not find container \"e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3\": container with ID starting with e460aae32a60d2a5a5ba81c1d108dd59ce1434ea768191ee34896936cc668be3 not found: ID does not exist" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.712841 4741 scope.go:117] "RemoveContainer" containerID="f74c820b505971e578bc8f81685750786c126b92ec697fda00bad07f4d638490" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.713274 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.720341 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.726986 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.735746 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-config-data" (OuterVolumeSpecName: "config-data") pod "93c5456d-7a59-4991-ad3d-58b04db78b24" (UID: "93c5456d-7a59-4991-ad3d-58b04db78b24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.741526 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi7d7d-account-delete-wklvh"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.775209 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6af0-account-delete-8hrbb"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.794756 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93c5456d-7a59-4991-ad3d-58b04db78b24" (UID: "93c5456d-7a59-4991-ad3d-58b04db78b24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.799752 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone81dc-account-delete-blpg9" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.805982 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-config-data\") pod \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806018 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-combined-ca-bundle\") pod \"2be2676f-d882-424e-8a49-d8bffc23773e\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806047 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-scripts\") pod \"2be2676f-d882-424e-8a49-d8bffc23773e\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806074 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-logs\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806101 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-scripts\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806132 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skkr5\" (UniqueName: \"kubernetes.io/projected/2be2676f-d882-424e-8a49-d8bffc23773e-kube-api-access-skkr5\") pod \"2be2676f-d882-424e-8a49-d8bffc23773e\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806162 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-internal-tls-certs\") pod \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806194 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-scripts\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806208 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2be2676f-d882-424e-8a49-d8bffc23773e-logs\") pod \"2be2676f-d882-424e-8a49-d8bffc23773e\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806236 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2tfl\" (UniqueName: \"kubernetes.io/projected/c9251106-ba81-446b-be7b-51c4ac7f3f81-kube-api-access-q2tfl\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806275 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-combined-ca-bundle\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806295 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-public-tls-certs\") pod \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806316 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-logs\") pod \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806342 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj2t8\" (UniqueName: \"kubernetes.io/projected/a0ade550-0eaf-4518-8696-14daf3f034d4-kube-api-access-nj2t8\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806369 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806424 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-config-data\") pod \"2be2676f-d882-424e-8a49-d8bffc23773e\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806470 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-internal-tls-certs\") pod \"2be2676f-d882-424e-8a49-d8bffc23773e\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806505 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-internal-tls-certs\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806524 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-httpd-run\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806539 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljjsz\" (UniqueName: \"kubernetes.io/projected/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-kube-api-access-ljjsz\") pod \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806557 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-config-data\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806586 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-logs\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806607 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-public-tls-certs\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806643 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-combined-ca-bundle\") pod \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\" (UID: \"be00d3f7-3dee-4e64-8559-2a5efe4ceba3\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806665 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-config-data\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806688 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-combined-ca-bundle\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806720 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"c9251106-ba81-446b-be7b-51c4ac7f3f81\" (UID: \"c9251106-ba81-446b-be7b-51c4ac7f3f81\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806738 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-httpd-run\") pod \"a0ade550-0eaf-4518-8696-14daf3f034d4\" (UID: \"a0ade550-0eaf-4518-8696-14daf3f034d4\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.806759 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-public-tls-certs\") pod \"2be2676f-d882-424e-8a49-d8bffc23773e\" (UID: \"2be2676f-d882-424e-8a49-d8bffc23773e\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.807134 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.807151 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c5456d-7a59-4991-ad3d-58b04db78b24-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.807160 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f8p4\" (UniqueName: \"kubernetes.io/projected/93c5456d-7a59-4991-ad3d-58b04db78b24-kube-api-access-6f8p4\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.816439 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9251106-ba81-446b-be7b-51c4ac7f3f81-kube-api-access-q2tfl" (OuterVolumeSpecName: "kube-api-access-q2tfl") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "kube-api-access-q2tfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.817786 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-logs" (OuterVolumeSpecName: "logs") pod "be00d3f7-3dee-4e64-8559-2a5efe4ceba3" (UID: "be00d3f7-3dee-4e64-8559-2a5efe4ceba3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.821066 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-logs" (OuterVolumeSpecName: "logs") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.823205 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-scripts" (OuterVolumeSpecName: "scripts") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.824332 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.830503 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.839669 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-scripts" (OuterVolumeSpecName: "scripts") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.840798 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2be2676f-d882-424e-8a49-d8bffc23773e-kube-api-access-skkr5" (OuterVolumeSpecName: "kube-api-access-skkr5") pod "2be2676f-d882-424e-8a49-d8bffc23773e" (UID: "2be2676f-d882-424e-8a49-d8bffc23773e"). InnerVolumeSpecName "kube-api-access-skkr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.841139 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2be2676f-d882-424e-8a49-d8bffc23773e-logs" (OuterVolumeSpecName: "logs") pod "2be2676f-d882-424e-8a49-d8bffc23773e" (UID: "2be2676f-d882-424e-8a49-d8bffc23773e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.844587 4741 scope.go:117] "RemoveContainer" containerID="e1a9547b01d309463905544f2e45064bcee20884139c4826affd282397c439fb" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.844961 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-logs" (OuterVolumeSpecName: "logs") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.855897 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.861317 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance6af0-account-delete-8hrbb"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.867702 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.868801 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.869173 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.872659 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.876957 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-kube-api-access-ljjsz" (OuterVolumeSpecName: "kube-api-access-ljjsz") pod "be00d3f7-3dee-4e64-8559-2a5efe4ceba3" (UID: "be00d3f7-3dee-4e64-8559-2a5efe4ceba3"). InnerVolumeSpecName "kube-api-access-ljjsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.883996 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.896688 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-scripts" (OuterVolumeSpecName: "scripts") pod "2be2676f-d882-424e-8a49-d8bffc23773e" (UID: "2be2676f-d882-424e-8a49-d8bffc23773e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.898786 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5c96956bc-7qhgv"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.909750 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ade550-0eaf-4518-8696-14daf3f034d4-kube-api-access-nj2t8" (OuterVolumeSpecName: "kube-api-access-nj2t8") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "kube-api-access-nj2t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910210 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2rwz\" (UniqueName: \"kubernetes.io/projected/6b0a22c3-d524-422d-8692-14d7a16a418f-kube-api-access-l2rwz\") pod \"6b0a22c3-d524-422d-8692-14d7a16a418f\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910274 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-internal-tls-certs\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910316 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-internal-tls-certs\") pod \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910359 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-public-tls-certs\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910384 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl6df\" (UniqueName: \"kubernetes.io/projected/e3b65efd-d659-4b2d-9cee-0c57b08d029f-kube-api-access-wl6df\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910427 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-certs\") pod \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910446 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-combined-ca-bundle\") pod \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910462 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data-custom\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910484 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910549 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldntx\" (UniqueName: \"kubernetes.io/projected/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-kube-api-access-ldntx\") pod \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910566 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-config\") pod \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910587 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbk2f\" (UniqueName: \"kubernetes.io/projected/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-api-access-wbk2f\") pod \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\" (UID: \"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910605 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3b65efd-d659-4b2d-9cee-0c57b08d029f-etc-machine-id\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910644 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-combined-ca-bundle\") pod \"6b0a22c3-d524-422d-8692-14d7a16a418f\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910687 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-config-data\") pod \"6b0a22c3-d524-422d-8692-14d7a16a418f\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910706 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data\") pod \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910724 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-combined-ca-bundle\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910757 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-nova-metadata-tls-certs\") pod \"6b0a22c3-d524-422d-8692-14d7a16a418f\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910774 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data-custom\") pod \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910797 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-scripts\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910833 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-combined-ca-bundle\") pod \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910851 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b0a22c3-d524-422d-8692-14d7a16a418f-logs\") pod \"6b0a22c3-d524-422d-8692-14d7a16a418f\" (UID: \"6b0a22c3-d524-422d-8692-14d7a16a418f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910874 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-public-tls-certs\") pod \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910894 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b65efd-d659-4b2d-9cee-0c57b08d029f-logs\") pod \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\" (UID: \"e3b65efd-d659-4b2d-9cee-0c57b08d029f\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.910909 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-logs\") pod \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\" (UID: \"58cfa8a3-5067-40ee-8e23-803a52d6b1b1\") " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911249 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911261 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911271 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911279 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911287 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911295 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skkr5\" (UniqueName: \"kubernetes.io/projected/2be2676f-d882-424e-8a49-d8bffc23773e-kube-api-access-skkr5\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911304 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911311 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2be2676f-d882-424e-8a49-d8bffc23773e-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911320 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2tfl\" (UniqueName: \"kubernetes.io/projected/c9251106-ba81-446b-be7b-51c4ac7f3f81-kube-api-access-q2tfl\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911328 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911337 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj2t8\" (UniqueName: \"kubernetes.io/projected/a0ade550-0eaf-4518-8696-14daf3f034d4-kube-api-access-nj2t8\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911351 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911360 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljjsz\" (UniqueName: \"kubernetes.io/projected/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-kube-api-access-ljjsz\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911368 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c9251106-ba81-446b-be7b-51c4ac7f3f81-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.911376 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ade550-0eaf-4518-8696-14daf3f034d4-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.921263 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b0a22c3-d524-422d-8692-14d7a16a418f-kube-api-access-l2rwz" (OuterVolumeSpecName: "kube-api-access-l2rwz") pod "6b0a22c3-d524-422d-8692-14d7a16a418f" (UID: "6b0a22c3-d524-422d-8692-14d7a16a418f"). InnerVolumeSpecName "kube-api-access-l2rwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.937134 4741 scope.go:117] "RemoveContainer" containerID="b14eefa9c9c65e040f41485625e023675c399fa3746763f7c230f6e4192d6b5e" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.938366 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b0a22c3-d524-422d-8692-14d7a16a418f-logs" (OuterVolumeSpecName: "logs") pod "6b0a22c3-d524-422d-8692-14d7a16a418f" (UID: "6b0a22c3-d524-422d-8692-14d7a16a418f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.939469 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3b65efd-d659-4b2d-9cee-0c57b08d029f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.940174 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-logs" (OuterVolumeSpecName: "logs") pod "58cfa8a3-5067-40ee-8e23-803a52d6b1b1" (UID: "58cfa8a3-5067-40ee-8e23-803a52d6b1b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.942725 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b65efd-d659-4b2d-9cee-0c57b08d029f-logs" (OuterVolumeSpecName: "logs") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.942813 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.944988 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-5c96956bc-7qhgv"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.947311 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "58cfa8a3-5067-40ee-8e23-803a52d6b1b1" (UID: "58cfa8a3-5067-40ee-8e23-803a52d6b1b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.951302 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-api-access-wbk2f" (OuterVolumeSpecName: "kube-api-access-wbk2f") pod "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" (UID: "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a"). InnerVolumeSpecName "kube-api-access-wbk2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.951416 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-kube-api-access-ldntx" (OuterVolumeSpecName: "kube-api-access-ldntx") pod "58cfa8a3-5067-40ee-8e23-803a52d6b1b1" (UID: "58cfa8a3-5067-40ee-8e23-803a52d6b1b1"). InnerVolumeSpecName "kube-api-access-ldntx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.958001 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-74d7c6f55-9trnx"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.967345 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-74d7c6f55-9trnx"] Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.983919 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b65efd-d659-4b2d-9cee-0c57b08d029f-kube-api-access-wl6df" (OuterVolumeSpecName: "kube-api-access-wl6df") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "kube-api-access-wl6df". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:09 crc kubenswrapper[4741]: I0929 19:31:09.984610 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-scripts" (OuterVolumeSpecName: "scripts") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.004987 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.008985 4741 scope.go:117] "RemoveContainer" containerID="5d7c9dcdb6e6631c524f68a99886213580f717eb067525088f01ff7e8451b3e4" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012711 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012735 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012747 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b0a22c3-d524-422d-8692-14d7a16a418f-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012756 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b65efd-d659-4b2d-9cee-0c57b08d029f-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012765 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-logs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012776 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2rwz\" (UniqueName: \"kubernetes.io/projected/6b0a22c3-d524-422d-8692-14d7a16a418f-kube-api-access-l2rwz\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012788 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012800 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl6df\" (UniqueName: \"kubernetes.io/projected/e3b65efd-d659-4b2d-9cee-0c57b08d029f-kube-api-access-wl6df\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012812 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012824 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldntx\" (UniqueName: \"kubernetes.io/projected/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-kube-api-access-ldntx\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012837 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbk2f\" (UniqueName: \"kubernetes.io/projected/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-api-access-wbk2f\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.012849 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3b65efd-d659-4b2d-9cee-0c57b08d029f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.098184 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-config-data" (OuterVolumeSpecName: "config-data") pod "be00d3f7-3dee-4e64-8559-2a5efe4ceba3" (UID: "be00d3f7-3dee-4e64-8559-2a5efe4ceba3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.099532 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell15b24-account-delete-whtzz" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.114077 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.144208 4741 scope.go:117] "RemoveContainer" containerID="f4540013c871a62255548905e7c7a5f8f143d8e4617fe18f7c6f7bdbd529b257" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.167442 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" (UID: "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.170516 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.170799 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.194534 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.216801 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bjgr\" (UniqueName: \"kubernetes.io/projected/13785b5b-f629-47ae-8251-331d44b40254-kube-api-access-7bjgr\") pod \"13785b5b-f629-47ae-8251-331d44b40254\" (UID: \"13785b5b-f629-47ae-8251-331d44b40254\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.218261 4741 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.218347 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.258879 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-8567fd5894-dwdtp"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.260992 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13785b5b-f629-47ae-8251-331d44b40254-kube-api-access-7bjgr" (OuterVolumeSpecName: "kube-api-access-7bjgr") pod "13785b5b-f629-47ae-8251-331d44b40254" (UID: "13785b5b-f629-47ae-8251-331d44b40254"). InnerVolumeSpecName "kube-api-access-7bjgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.270079 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-8567fd5894-dwdtp"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.273816 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-config-data" (OuterVolumeSpecName: "config-data") pod "6b0a22c3-d524-422d-8692-14d7a16a418f" (UID: "6b0a22c3-d524-422d-8692-14d7a16a418f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.306035 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.319506 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e0c02dc_69a9_4e60_b179_0e23842d10a4.slice/crio-conmon-7d7103d33c61d8d8489afa8a3b1ea5db27133d9afa53b11454cfade2f1762a5c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e0c02dc_69a9_4e60_b179_0e23842d10a4.slice/crio-7d7103d33c61d8d8489afa8a3b1ea5db27133d9afa53b11454cfade2f1762a5c.scope\": RecentStats: unable to find data in memory cache]" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.319704 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b0a22c3-d524-422d-8692-14d7a16a418f" (UID: "6b0a22c3-d524-422d-8692-14d7a16a418f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.320353 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.320418 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.320429 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.320439 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bjgr\" (UniqueName: \"kubernetes.io/projected/13785b5b-f629-47ae-8251-331d44b40254-kube-api-access-7bjgr\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.360200 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.362712 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.366735 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.369086 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.369168 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="cb846a7f-9be5-4a52-837e-a423f7736e79" containerName="nova-cell0-conductor-conductor" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.390114 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.423777 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.423818 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.464518 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be00d3f7-3dee-4e64-8559-2a5efe4ceba3" (UID: "be00d3f7-3dee-4e64-8559-2a5efe4ceba3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.476565 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "58cfa8a3-5067-40ee-8e23-803a52d6b1b1" (UID: "58cfa8a3-5067-40ee-8e23-803a52d6b1b1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.497437 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" (UID: "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.505577 4741 generic.go:334] "Generic (PLEG): container finished" podID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerID="7d7103d33c61d8d8489afa8a3b1ea5db27133d9afa53b11454cfade2f1762a5c" exitCode=0 Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.505642 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8e0c02dc-69a9-4e60-b179-0e23842d10a4","Type":"ContainerDied","Data":"7d7103d33c61d8d8489afa8a3b1ea5db27133d9afa53b11454cfade2f1762a5c"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.510050 4741 generic.go:334] "Generic (PLEG): container finished" podID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerID="05c981478a21299cdffcc57230be7582dcc6e046d2f10f197e1a1fb4606f05e6" exitCode=0 Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.510125 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df6beb49-03ad-47ef-a9c7-3f37baa6d105","Type":"ContainerDied","Data":"05c981478a21299cdffcc57230be7582dcc6e046d2f10f197e1a1fb4606f05e6"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.513200 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell15b24-account-delete-whtzz" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.513224 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell15b24-account-delete-whtzz" event={"ID":"13785b5b-f629-47ae-8251-331d44b40254","Type":"ContainerDied","Data":"b620ebe1c97698a6498d508a46e150f1702c9557a9a398daa2eb00b541592b9f"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.513286 4741 scope.go:117] "RemoveContainer" containerID="fe1574bb5521288388babc106992a52d9ec3cca29d6d836f9f0d16b9960f0890" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.518755 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.519546 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9d4bbc33-c647-4eda-bdc6-89f80678ae8d/ovn-northd/0.log" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.519600 4741 generic.go:334] "Generic (PLEG): container finished" podID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerID="3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96" exitCode=139 Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.519671 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9d4bbc33-c647-4eda-bdc6-89f80678ae8d","Type":"ContainerDied","Data":"3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.520953 4741 generic.go:334] "Generic (PLEG): container finished" podID="9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" containerID="7c71b4d390feacc6365d3ebfe09f6e223b707ee503a77ad41f8f50b865956428" exitCode=0 Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.521008 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3","Type":"ContainerDied","Data":"7c71b4d390feacc6365d3ebfe09f6e223b707ee503a77ad41f8f50b865956428"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.521025 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3","Type":"ContainerDied","Data":"954aa19a272d9b6aad0c25106b859c57b0a12fd3fd572bba3b9fe7fef141343b"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.521035 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="954aa19a272d9b6aad0c25106b859c57b0a12fd3fd572bba3b9fe7fef141343b" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.521591 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.522112 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5c39883d-b9e0-4b1f-a7b0-8d29c04d066a","Type":"ContainerDied","Data":"bdb6ca44925061fd0dc3065188cdf0b6fadd4d39e0bf0c0bf750a7c29bcc594c"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.522224 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.524684 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.524706 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.524716 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.524726 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.524735 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.527676 4741 generic.go:334] "Generic (PLEG): container finished" podID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerID="64aaaacdbcb9f52035038a67edb1624996ab0fb313a96a375112066aa4d0c3a6" exitCode=0 Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.527772 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.527802 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.527830 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerDied","Data":"64aaaacdbcb9f52035038a67edb1624996ab0fb313a96a375112066aa4d0c3a6"} Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.527848 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.528091 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.528163 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.528212 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.528237 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c686d89f8-xwpjv" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.528348 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659f5d9f44-j8zgh" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.528453 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone81dc-account-delete-blpg9" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.534334 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "58cfa8a3-5067-40ee-8e23-803a52d6b1b1" (UID: "58cfa8a3-5067-40ee-8e23-803a52d6b1b1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.534619 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be00d3f7-3dee-4e64-8559-2a5efe4ceba3" (UID: "be00d3f7-3dee-4e64-8559-2a5efe4ceba3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.572479 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.573125 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58cfa8a3-5067-40ee-8e23-803a52d6b1b1" (UID: "58cfa8a3-5067-40ee-8e23-803a52d6b1b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.594572 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6b0a22c3-d524-422d-8692-14d7a16a418f" (UID: "6b0a22c3-d524-422d-8692-14d7a16a418f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.597326 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" (UID: "5c39883d-b9e0-4b1f-a7b0-8d29c04d066a"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.628506 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.628508 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.628548 4741 secret.go:188] Couldn't get secret openstack/keystone-config-data: secret "keystone-config-data" not found Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.628694 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.628755 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:12.628734729 +0000 UTC m=+1314.276524061 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-config-data" not found Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.628900 4741 secret.go:188] Couldn't get secret openstack/keystone-scripts: secret "keystone-scripts" not found Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.628980 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:12.628959436 +0000 UTC m=+1314.276748768 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-scripts" not found Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.629010 4741 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.629037 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:12.629032038 +0000 UTC m=+1314.276821370 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.629048 4741 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0a22c3-d524-422d-8692-14d7a16a418f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.629063 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.629074 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.629101 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:10 crc kubenswrapper[4741]: E0929 19:31:10.629131 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:12.629123831 +0000 UTC m=+1314.276913163 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.632957 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data" (OuterVolumeSpecName: "config-data") pod "58cfa8a3-5067-40ee-8e23-803a52d6b1b1" (UID: "58cfa8a3-5067-40ee-8e23-803a52d6b1b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.637791 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-config-data" (OuterVolumeSpecName: "config-data") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.640380 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data" (OuterVolumeSpecName: "config-data") pod "e3b65efd-d659-4b2d-9cee-0c57b08d029f" (UID: "e3b65efd-d659-4b2d-9cee-0c57b08d029f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.674544 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c9251106-ba81-446b-be7b-51c4ac7f3f81" (UID: "c9251106-ba81-446b-be7b-51c4ac7f3f81"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.676665 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2be2676f-d882-424e-8a49-d8bffc23773e" (UID: "2be2676f-d882-424e-8a49-d8bffc23773e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.677882 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-config-data" (OuterVolumeSpecName: "config-data") pod "2be2676f-d882-424e-8a49-d8bffc23773e" (UID: "2be2676f-d882-424e-8a49-d8bffc23773e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.682759 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be00d3f7-3dee-4e64-8559-2a5efe4ceba3" (UID: "be00d3f7-3dee-4e64-8559-2a5efe4ceba3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.686060 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2be2676f-d882-424e-8a49-d8bffc23773e" (UID: "2be2676f-d882-424e-8a49-d8bffc23773e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.690226 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-config-data" (OuterVolumeSpecName: "config-data") pod "a0ade550-0eaf-4518-8696-14daf3f034d4" (UID: "a0ade550-0eaf-4518-8696-14daf3f034d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.707841 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2be2676f-d882-424e-8a49-d8bffc23773e" (UID: "2be2676f-d882-424e-8a49-d8bffc23773e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.725756 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730133 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730162 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730171 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730180 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9251106-ba81-446b-be7b-51c4ac7f3f81-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730188 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b65efd-d659-4b2d-9cee-0c57b08d029f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730196 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ade550-0eaf-4518-8696-14daf3f034d4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730204 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730212 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2676f-d882-424e-8a49-d8bffc23773e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730221 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cfa8a3-5067-40ee-8e23-803a52d6b1b1-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.730231 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be00d3f7-3dee-4e64-8559-2a5efe4ceba3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.731318 4741 scope.go:117] "RemoveContainer" containerID="7720681edbf8321fa29bc81ceddd38f098ef1fa29db00e81fd9806812c014bd7" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.748495 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.782378 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.794220 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9d4bbc33-c647-4eda-bdc6-89f80678ae8d/ovn-northd/0.log" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.794305 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.804122 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell15b24-account-delete-whtzz"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.819947 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell15b24-account-delete-whtzz"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.832452 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.838730 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.838900 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.838944 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-config-data\") pod \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.838973 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-combined-ca-bundle\") pod \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839038 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-rundir\") pod \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839095 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlwcm\" (UniqueName: \"kubernetes.io/projected/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kube-api-access-jlwcm\") pod \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839145 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-erlang-cookie\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839207 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cj96\" (UniqueName: \"kubernetes.io/projected/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-kube-api-access-2cj96\") pod \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839258 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-tls\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839282 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-server-conf\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839303 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-scripts\") pod \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839331 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-plugins-conf\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839361 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839399 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df6beb49-03ad-47ef-a9c7-3f37baa6d105-pod-info\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839417 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-confd\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839433 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kolla-config\") pod \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839458 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-erlang-cookie\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839553 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-confd\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839575 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-combined-ca-bundle\") pod \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839601 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-tls\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839617 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-plugins\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839633 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-plugins-conf\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839647 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839671 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e0c02dc-69a9-4e60-b179-0e23842d10a4-pod-info\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839685 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839704 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-server-conf\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839728 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e0c02dc-69a9-4e60-b179-0e23842d10a4-erlang-cookie-secret\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839769 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-memcached-tls-certs\") pod \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\" (UID: \"9e8cf4ba-5168-48b5-8424-3e3e21bbdae3\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839785 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-northd-tls-certs\") pod \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839807 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q959n\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-kube-api-access-q959n\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839823 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-config\") pod \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839843 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-plugins\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839859 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxgqj\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-kube-api-access-jxgqj\") pod \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\" (UID: \"8e0c02dc-69a9-4e60-b179-0e23842d10a4\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839883 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df6beb49-03ad-47ef-a9c7-3f37baa6d105-erlang-cookie-secret\") pod \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\" (UID: \"df6beb49-03ad-47ef-a9c7-3f37baa6d105\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.839903 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-metrics-certs-tls-certs\") pod \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\" (UID: \"9d4bbc33-c647-4eda-bdc6-89f80678ae8d\") " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.841478 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "9d4bbc33-c647-4eda-bdc6-89f80678ae8d" (UID: "9d4bbc33-c647-4eda-bdc6-89f80678ae8d"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.844735 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" (UID: "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.845405 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-config" (OuterVolumeSpecName: "config") pod "9d4bbc33-c647-4eda-bdc6-89f80678ae8d" (UID: "9d4bbc33-c647-4eda-bdc6-89f80678ae8d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.845701 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.846038 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.846192 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-config-data" (OuterVolumeSpecName: "config-data") pod "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" (UID: "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.847211 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.847997 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.848075 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/df6beb49-03ad-47ef-a9c7-3f37baa6d105-pod-info" (OuterVolumeSpecName: "pod-info") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.848528 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-scripts" (OuterVolumeSpecName: "scripts") pod "9d4bbc33-c647-4eda-bdc6-89f80678ae8d" (UID: "9d4bbc33-c647-4eda-bdc6-89f80678ae8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.851592 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.855875 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.868498 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone81dc-account-delete-blpg9"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.872668 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6beb49-03ad-47ef-a9c7-3f37baa6d105-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.884467 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone81dc-account-delete-blpg9"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.900253 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-kube-api-access-q959n" (OuterVolumeSpecName: "kube-api-access-q959n") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "kube-api-access-q959n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.900530 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.900778 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.901539 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-kube-api-access-2cj96" (OuterVolumeSpecName: "kube-api-access-2cj96") pod "9d4bbc33-c647-4eda-bdc6-89f80678ae8d" (UID: "9d4bbc33-c647-4eda-bdc6-89f80678ae8d"). InnerVolumeSpecName "kube-api-access-2cj96". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.903665 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-kube-api-access-jxgqj" (OuterVolumeSpecName: "kube-api-access-jxgqj") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "kube-api-access-jxgqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.908737 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.919565 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.926984 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kube-api-access-jlwcm" (OuterVolumeSpecName: "kube-api-access-jlwcm") pod "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" (UID: "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3"). InnerVolumeSpecName "kube-api-access-jlwcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.927103 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8e0c02dc-69a9-4e60-b179-0e23842d10a4-pod-info" (OuterVolumeSpecName: "pod-info") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.928178 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0c02dc-69a9-4e60-b179-0e23842d10a4-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.928842 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data" (OuterVolumeSpecName: "config-data") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.933138 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data" (OuterVolumeSpecName: "config-data") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957051 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957085 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957095 4741 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957105 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957115 4741 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df6beb49-03ad-47ef-a9c7-3f37baa6d105-pod-info\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957123 4741 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kolla-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957133 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957142 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957151 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957160 4741 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957180 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957189 4741 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e0c02dc-69a9-4e60-b179-0e23842d10a4-pod-info\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957202 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957210 4741 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e0c02dc-69a9-4e60-b179-0e23842d10a4-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957219 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q959n\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-kube-api-access-q959n\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957230 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957238 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957247 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxgqj\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-kube-api-access-jxgqj\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957256 4741 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df6beb49-03ad-47ef-a9c7-3f37baa6d105-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957264 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957272 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957279 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-rundir\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957290 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlwcm\" (UniqueName: \"kubernetes.io/projected/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-kube-api-access-jlwcm\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957298 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.957307 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cj96\" (UniqueName: \"kubernetes.io/projected/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-kube-api-access-2cj96\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.970480 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:31:10 crc kubenswrapper[4741]: I0929 19:31:10.975912 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d4bbc33-c647-4eda-bdc6-89f80678ae8d" (UID: "9d4bbc33-c647-4eda-bdc6-89f80678ae8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.007020 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.012583 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.023456 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.036000 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" (UID: "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.038061 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.039077 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.053661 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "9d4bbc33-c647-4eda-bdc6-89f80678ae8d" (UID: "9d4bbc33-c647-4eda-bdc6-89f80678ae8d"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.059697 4741 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.059745 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.059754 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.059769 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.059779 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.062945 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.063642 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" (UID: "9e8cf4ba-5168-48b5-8424-3e3e21bbdae3"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.075368 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-server-conf" (OuterVolumeSpecName: "server-conf") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.106041 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-server-conf" (OuterVolumeSpecName: "server-conf") pod "8e0c02dc-69a9-4e60-b179-0e23842d10a4" (UID: "8e0c02dc-69a9-4e60-b179-0e23842d10a4"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.117079 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13785b5b-f629-47ae-8251-331d44b40254" path="/var/lib/kubelet/pods/13785b5b-f629-47ae-8251-331d44b40254/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.121923 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" path="/var/lib/kubelet/pods/2efb9fff-2e44-4257-bbe6-e5dccedcdcbe/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.122621 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" path="/var/lib/kubelet/pods/5c39883d-b9e0-4b1f-a7b0-8d29c04d066a/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.122747 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "9d4bbc33-c647-4eda-bdc6-89f80678ae8d" (UID: "9d4bbc33-c647-4eda-bdc6-89f80678ae8d"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.123591 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c5456d-7a59-4991-ad3d-58b04db78b24" path="/var/lib/kubelet/pods/93c5456d-7a59-4991-ad3d-58b04db78b24/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.124025 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="983d2897-24d2-462c-b9f9-427639d2b8f9" path="/var/lib/kubelet/pods/983d2897-24d2-462c-b9f9-427639d2b8f9/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.124479 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d18b4c1-d423-4840-97ff-c322272c3aa3" path="/var/lib/kubelet/pods/9d18b4c1-d423-4840-97ff-c322272c3aa3/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.124871 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62" path="/var/lib/kubelet/pods/9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.126077 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" path="/var/lib/kubelet/pods/a0ade550-0eaf-4518-8696-14daf3f034d4/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.126790 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8c0591c-26f3-48eb-9553-e61c932fda9d" path="/var/lib/kubelet/pods/b8c0591c-26f3-48eb-9553-e61c932fda9d/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.127246 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" path="/var/lib/kubelet/pods/c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.129997 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" path="/var/lib/kubelet/pods/d3e3a43d-5ff8-491d-916c-3c0b9e31f223/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.131164 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b64f81-4c6d-4d71-8930-109240d8ec42" path="/var/lib/kubelet/pods/e3b64f81-4c6d-4d71-8930-109240d8ec42/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.132286 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f24ec6be-6832-4eee-916c-2237e81bf0f1" path="/var/lib/kubelet/pods/f24ec6be-6832-4eee-916c-2237e81bf0f1/volumes" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.141742 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.141781 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.146121 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c686d89f8-xwpjv"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.146694 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "df6beb49-03ad-47ef-a9c7-3f37baa6d105" (UID: "df6beb49-03ad-47ef-a9c7-3f37baa6d105"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.154346 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c686d89f8-xwpjv"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.161050 4741 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e0c02dc-69a9-4e60-b179-0e23842d10a4-server-conf\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.161075 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e0c02dc-69a9-4e60-b179-0e23842d10a4-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.161100 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df6beb49-03ad-47ef-a9c7-3f37baa6d105-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.161110 4741 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df6beb49-03ad-47ef-a9c7-3f37baa6d105-server-conf\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.161118 4741 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.161127 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4bbc33-c647-4eda-bdc6-89f80678ae8d-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.166683 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.173751 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.179484 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.183838 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.188458 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.193680 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.209046 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-659f5d9f44-j8zgh"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.209108 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-659f5d9f44-j8zgh"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.460611 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.537897 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df6beb49-03ad-47ef-a9c7-3f37baa6d105","Type":"ContainerDied","Data":"52323e4c4ff52533356a2be38ec41b4ab9e9b44d885f039a3210f5f905046db6"} Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.537924 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.537942 4741 scope.go:117] "RemoveContainer" containerID="05c981478a21299cdffcc57230be7582dcc6e046d2f10f197e1a1fb4606f05e6" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.540897 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9d4bbc33-c647-4eda-bdc6-89f80678ae8d/ovn-northd/0.log" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.540963 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9d4bbc33-c647-4eda-bdc6-89f80678ae8d","Type":"ContainerDied","Data":"7a4cadaa08ec960786e08317688380e6c5819fa220f871529e224f3445c214e1"} Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.541058 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.547374 4741 generic.go:334] "Generic (PLEG): container finished" podID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerID="535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173" exitCode=0 Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.547415 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fad684df-76f3-4651-81e8-9bb739cfbc4b","Type":"ContainerDied","Data":"535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173"} Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.547444 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fad684df-76f3-4651-81e8-9bb739cfbc4b","Type":"ContainerDied","Data":"2c2c11f363480f2bcbf633c4aec8f098841d5ce4c315bb29697fdb705aecd9cc"} Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.547444 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.549743 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8e0c02dc-69a9-4e60-b179-0e23842d10a4","Type":"ContainerDied","Data":"80a92dcabd9ec3b723fea5b8cfae1457d22e984d367ebefb9b977d9fef4a156c"} Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.549780 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.550759 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566140 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-operator-scripts\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566197 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc5gs\" (UniqueName: \"kubernetes.io/projected/fad684df-76f3-4651-81e8-9bb739cfbc4b-kube-api-access-wc5gs\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566290 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-kolla-config\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566319 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-galera-tls-certs\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566341 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-secrets\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566375 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-combined-ca-bundle\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566452 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-default\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566483 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.566501 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-generated\") pod \"fad684df-76f3-4651-81e8-9bb739cfbc4b\" (UID: \"fad684df-76f3-4651-81e8-9bb739cfbc4b\") " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.567534 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.568309 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.573859 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.574124 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.579926 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fad684df-76f3-4651-81e8-9bb739cfbc4b-kube-api-access-wc5gs" (OuterVolumeSpecName: "kube-api-access-wc5gs") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "kube-api-access-wc5gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.588312 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-secrets" (OuterVolumeSpecName: "secrets") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.588435 4741 scope.go:117] "RemoveContainer" containerID="298a19b9e2f8e061524da9a8a4b1326a3050a055dc372e86be58a14f3e42e0e3" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.591206 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.606153 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.653740 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "fad684df-76f3-4651-81e8-9bb739cfbc4b" (UID: "fad684df-76f3-4651-81e8-9bb739cfbc4b"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668553 4741 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668589 4741 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668601 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad684df-76f3-4651-81e8-9bb739cfbc4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668613 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-default\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668648 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668660 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fad684df-76f3-4651-81e8-9bb739cfbc4b-config-data-generated\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668674 4741 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-operator-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668685 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc5gs\" (UniqueName: \"kubernetes.io/projected/fad684df-76f3-4651-81e8-9bb739cfbc4b-kube-api-access-wc5gs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.668696 4741 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fad684df-76f3-4651-81e8-9bb739cfbc4b-kolla-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.686357 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.735650 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.745448 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.751194 4741 scope.go:117] "RemoveContainer" containerID="e2974921beee8fdb66bfc522f1bf2e042117a87afcc7cd9f689c79ab47e6bcf2" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.758547 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.769093 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.769765 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.777031 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.778899 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.783600 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.787582 4741 scope.go:117] "RemoveContainer" containerID="3c423746198790a04f0c59460acfca51363f87f114c3aae39fa97f64a1f67a96" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.787869 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.804667 4741 scope.go:117] "RemoveContainer" containerID="535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.824589 4741 scope.go:117] "RemoveContainer" containerID="6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.852721 4741 scope.go:117] "RemoveContainer" containerID="535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173" Sep 29 19:31:11 crc kubenswrapper[4741]: E0929 19:31:11.853224 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173\": container with ID starting with 535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173 not found: ID does not exist" containerID="535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.853276 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173"} err="failed to get container status \"535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173\": rpc error: code = NotFound desc = could not find container \"535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173\": container with ID starting with 535c5750f02e3bd546cf3e7629e97761c95da2b0a7e0918523c2396e42f52173 not found: ID does not exist" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.853310 4741 scope.go:117] "RemoveContainer" containerID="6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311" Sep 29 19:31:11 crc kubenswrapper[4741]: E0929 19:31:11.855357 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311\": container with ID starting with 6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311 not found: ID does not exist" containerID="6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.855400 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311"} err="failed to get container status \"6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311\": rpc error: code = NotFound desc = could not find container \"6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311\": container with ID starting with 6d3d053a7c014f7e078dcb4cc7eec0afb285280c12dcb4f9b2cd5889ec876311 not found: ID does not exist" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.855424 4741 scope.go:117] "RemoveContainer" containerID="7d7103d33c61d8d8489afa8a3b1ea5db27133d9afa53b11454cfade2f1762a5c" Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.887977 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.893042 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 19:31:11 crc kubenswrapper[4741]: I0929 19:31:11.896849 4741 scope.go:117] "RemoveContainer" containerID="7b27d2ea63000a29d7fab7a6fb5ae477f330f125488a691cbc9155958f46378e" Sep 29 19:31:11 crc kubenswrapper[4741]: E0929 19:31:11.972599 4741 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Sep 29 19:31:11 crc kubenswrapper[4741]: E0929 19:31:11.972664 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:19.972649886 +0000 UTC m=+1321.620439218 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-config" not found Sep 29 19:31:11 crc kubenswrapper[4741]: E0929 19:31:11.972598 4741 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Sep 29 19:31:11 crc kubenswrapper[4741]: E0929 19:31:11.972786 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:19.97276459 +0000 UTC m=+1321.620553922 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-httpd-config" not found Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.013492 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.039993 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.072963 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-scripts\") pod \"c907b3ee-67ab-4d29-a02d-96137c124ff4\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.073017 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-combined-ca-bundle\") pod \"c907b3ee-67ab-4d29-a02d-96137c124ff4\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.073041 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dqxx\" (UniqueName: \"kubernetes.io/projected/c907b3ee-67ab-4d29-a02d-96137c124ff4-kube-api-access-2dqxx\") pod \"c907b3ee-67ab-4d29-a02d-96137c124ff4\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.073067 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data\") pod \"c907b3ee-67ab-4d29-a02d-96137c124ff4\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.073087 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c907b3ee-67ab-4d29-a02d-96137c124ff4-etc-machine-id\") pod \"c907b3ee-67ab-4d29-a02d-96137c124ff4\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.073122 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data-custom\") pod \"c907b3ee-67ab-4d29-a02d-96137c124ff4\" (UID: \"c907b3ee-67ab-4d29-a02d-96137c124ff4\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.075879 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c907b3ee-67ab-4d29-a02d-96137c124ff4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c907b3ee-67ab-4d29-a02d-96137c124ff4" (UID: "c907b3ee-67ab-4d29-a02d-96137c124ff4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.077567 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c907b3ee-67ab-4d29-a02d-96137c124ff4" (UID: "c907b3ee-67ab-4d29-a02d-96137c124ff4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.077591 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c907b3ee-67ab-4d29-a02d-96137c124ff4-kube-api-access-2dqxx" (OuterVolumeSpecName: "kube-api-access-2dqxx") pod "c907b3ee-67ab-4d29-a02d-96137c124ff4" (UID: "c907b3ee-67ab-4d29-a02d-96137c124ff4"). InnerVolumeSpecName "kube-api-access-2dqxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.077639 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-scripts" (OuterVolumeSpecName: "scripts") pod "c907b3ee-67ab-4d29-a02d-96137c124ff4" (UID: "c907b3ee-67ab-4d29-a02d-96137c124ff4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.111538 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c907b3ee-67ab-4d29-a02d-96137c124ff4" (UID: "c907b3ee-67ab-4d29-a02d-96137c124ff4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.153027 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data" (OuterVolumeSpecName: "config-data") pod "c907b3ee-67ab-4d29-a02d-96137c124ff4" (UID: "c907b3ee-67ab-4d29-a02d-96137c124ff4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.174871 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-combined-ca-bundle\") pod \"cb846a7f-9be5-4a52-837e-a423f7736e79\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.174985 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lqb9\" (UniqueName: \"kubernetes.io/projected/cb846a7f-9be5-4a52-837e-a423f7736e79-kube-api-access-8lqb9\") pod \"cb846a7f-9be5-4a52-837e-a423f7736e79\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.175069 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-config-data\") pod \"cb846a7f-9be5-4a52-837e-a423f7736e79\" (UID: \"cb846a7f-9be5-4a52-837e-a423f7736e79\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.177104 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c907b3ee-67ab-4d29-a02d-96137c124ff4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.177128 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.177140 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.177151 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.177163 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dqxx\" (UniqueName: \"kubernetes.io/projected/c907b3ee-67ab-4d29-a02d-96137c124ff4-kube-api-access-2dqxx\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.177175 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c907b3ee-67ab-4d29-a02d-96137c124ff4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.180866 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb846a7f-9be5-4a52-837e-a423f7736e79-kube-api-access-8lqb9" (OuterVolumeSpecName: "kube-api-access-8lqb9") pod "cb846a7f-9be5-4a52-837e-a423f7736e79" (UID: "cb846a7f-9be5-4a52-837e-a423f7736e79"). InnerVolumeSpecName "kube-api-access-8lqb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.199748 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-config-data" (OuterVolumeSpecName: "config-data") pod "cb846a7f-9be5-4a52-837e-a423f7736e79" (UID: "cb846a7f-9be5-4a52-837e-a423f7736e79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.200445 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb846a7f-9be5-4a52-837e-a423f7736e79" (UID: "cb846a7f-9be5-4a52-837e-a423f7736e79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.278508 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.278542 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb846a7f-9be5-4a52-837e-a423f7736e79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.278556 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lqb9\" (UniqueName: \"kubernetes.io/projected/cb846a7f-9be5-4a52-837e-a423f7736e79-kube-api-access-8lqb9\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.579447 4741 generic.go:334] "Generic (PLEG): container finished" podID="cb846a7f-9be5-4a52-837e-a423f7736e79" containerID="e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" exitCode=0 Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.579565 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cb846a7f-9be5-4a52-837e-a423f7736e79","Type":"ContainerDied","Data":"e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144"} Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.580352 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cb846a7f-9be5-4a52-837e-a423f7736e79","Type":"ContainerDied","Data":"179d9546cedc9ade539f8578cc2b0182f784899b30f3f6e0935dd69d19864d5d"} Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.580375 4741 scope.go:117] "RemoveContainer" containerID="e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.579555 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.605292 4741 generic.go:334] "Generic (PLEG): container finished" podID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerID="a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04" exitCode=0 Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.605353 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c907b3ee-67ab-4d29-a02d-96137c124ff4","Type":"ContainerDied","Data":"a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04"} Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.605378 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c907b3ee-67ab-4d29-a02d-96137c124ff4","Type":"ContainerDied","Data":"233545f1b736491909602c4210ef4aa4ef9018b304a5faf889117f508dafcc56"} Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.605374 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.617799 4741 generic.go:334] "Generic (PLEG): container finished" podID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerID="3d41a9cc13ba07eab14e449222ca1eba4bc2f913c9d707b7f2f9fe2953779503" exitCode=0 Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.617894 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerDied","Data":"3d41a9cc13ba07eab14e449222ca1eba4bc2f913c9d707b7f2f9fe2953779503"} Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.624150 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.625364 4741 scope.go:117] "RemoveContainer" containerID="e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.625968 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144\": container with ID starting with e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144 not found: ID does not exist" containerID="e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.626007 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144"} err="failed to get container status \"e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144\": rpc error: code = NotFound desc = could not find container \"e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144\": container with ID starting with e98b992971f0f02d03988a48bdc8436f8b2cdc427b1563ae8b9945b6ca148144 not found: ID does not exist" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.626031 4741 scope.go:117] "RemoveContainer" containerID="5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.654287 4741 scope.go:117] "RemoveContainer" containerID="a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.655570 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.660536 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.664919 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.683513 4741 scope.go:117] "RemoveContainer" containerID="5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7" Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.683970 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7\": container with ID starting with 5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7 not found: ID does not exist" containerID="5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.684003 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7"} err="failed to get container status \"5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7\": rpc error: code = NotFound desc = could not find container \"5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7\": container with ID starting with 5cc2049eaed03c9ab5fe473dba6b0bee3432999f71ad9db39b4022e88514b2f7 not found: ID does not exist" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.684027 4741 scope.go:117] "RemoveContainer" containerID="a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04" Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.684241 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04\": container with ID starting with a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04 not found: ID does not exist" containerID="a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.684266 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04"} err="failed to get container status \"a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04\": rpc error: code = NotFound desc = could not find container \"a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04\": container with ID starting with a06b358df3fb8f1da445019f433e7e05e139c6d3d2fef0f4ce0609c152222c04 not found: ID does not exist" Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685234 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685243 4741 secret.go:188] Couldn't get secret openstack/keystone-config-data: secret "keystone-config-data" not found Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685314 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:16.685296385 +0000 UTC m=+1318.333085717 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685331 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:16.685324516 +0000 UTC m=+1318.333113848 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-config-data" not found Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685417 4741 secret.go:188] Couldn't get secret openstack/keystone-scripts: secret "keystone-scripts" not found Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685483 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:16.68546784 +0000 UTC m=+1318.333257162 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone-scripts" not found Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685562 4741 secret.go:188] Couldn't get secret openstack/keystone: secret "keystone" not found Sep 29 19:31:12 crc kubenswrapper[4741]: E0929 19:31:12.685595 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys podName:c81db200-8fff-4a29-952c-4f8609dc8ac8 nodeName:}" failed. No retries permitted until 2025-09-29 19:31:16.685586734 +0000 UTC m=+1318.333376066 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys") pod "keystone-677b8f6f88-c7kf7" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8") : secret "keystone" not found Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.866047 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.982546 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988263 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-config-data\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988330 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-combined-ca-bundle\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988488 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhs7q\" (UniqueName: \"kubernetes.io/projected/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-kube-api-access-hhs7q\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988527 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-log-httpd\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988582 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-run-httpd\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988598 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-sg-core-conf-yaml\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988631 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-ceilometer-tls-certs\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.988663 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-scripts\") pod \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\" (UID: \"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834\") " Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.989508 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.989419 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.992699 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-kube-api-access-hhs7q" (OuterVolumeSpecName: "kube-api-access-hhs7q") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "kube-api-access-hhs7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:12 crc kubenswrapper[4741]: I0929 19:31:12.993314 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-scripts" (OuterVolumeSpecName: "scripts") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.020580 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.033190 4741 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Sep 29 19:31:13 crc kubenswrapper[4741]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-09-29T19:31:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Sep 29 19:31:13 crc kubenswrapper[4741]: /etc/init.d/functions: line 589: 365 Alarm clock "$@" Sep 29 19:31:13 crc kubenswrapper[4741]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-68snt" message=< Sep 29 19:31:13 crc kubenswrapper[4741]: Exiting ovn-controller (1) [FAILED] Sep 29 19:31:13 crc kubenswrapper[4741]: Killing ovn-controller (1) [ OK ] Sep 29 19:31:13 crc kubenswrapper[4741]: Killing ovn-controller (1) with SIGKILL [ OK ] Sep 29 19:31:13 crc kubenswrapper[4741]: 2025-09-29T19:31:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Sep 29 19:31:13 crc kubenswrapper[4741]: /etc/init.d/functions: line 589: 365 Alarm clock "$@" Sep 29 19:31:13 crc kubenswrapper[4741]: > Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.033242 4741 kuberuntime_container.go:691] "PreStop hook failed" err=< Sep 29 19:31:13 crc kubenswrapper[4741]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-09-29T19:31:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Sep 29 19:31:13 crc kubenswrapper[4741]: /etc/init.d/functions: line 589: 365 Alarm clock "$@" Sep 29 19:31:13 crc kubenswrapper[4741]: > pod="openstack/ovn-controller-68snt" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerName="ovn-controller" containerID="cri-o://1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.033292 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-68snt" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerName="ovn-controller" containerID="cri-o://1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0" gracePeriod=22 Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.045235 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.066781 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089216 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089265 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089318 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-internal-tls-certs\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089337 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-public-tls-certs\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089473 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089498 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-combined-ca-bundle\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089527 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.089564 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hklql\" (UniqueName: \"kubernetes.io/projected/c81db200-8fff-4a29-952c-4f8609dc8ac8-kube-api-access-hklql\") pod \"c81db200-8fff-4a29-952c-4f8609dc8ac8\" (UID: \"c81db200-8fff-4a29-952c-4f8609dc8ac8\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.090139 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhs7q\" (UniqueName: \"kubernetes.io/projected/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-kube-api-access-hhs7q\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.090155 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.090165 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.090174 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.090184 4741 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.090192 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.090200 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.094701 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.094768 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts" (OuterVolumeSpecName: "scripts") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.095326 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.097597 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" path="/var/lib/kubelet/pods/2be2676f-d882-424e-8a49-d8bffc23773e/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.097906 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c81db200-8fff-4a29-952c-4f8609dc8ac8-kube-api-access-hklql" (OuterVolumeSpecName: "kube-api-access-hklql") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "kube-api-access-hklql". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.098200 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" path="/var/lib/kubelet/pods/58cfa8a3-5067-40ee-8e23-803a52d6b1b1/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.098802 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" path="/var/lib/kubelet/pods/6b0a22c3-d524-422d-8692-14d7a16a418f/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.099934 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" path="/var/lib/kubelet/pods/8e0c02dc-69a9-4e60-b179-0e23842d10a4/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.100533 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" path="/var/lib/kubelet/pods/9d4bbc33-c647-4eda-bdc6-89f80678ae8d/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.101639 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" path="/var/lib/kubelet/pods/9e8cf4ba-5168-48b5-8424-3e3e21bbdae3/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.102100 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" path="/var/lib/kubelet/pods/be00d3f7-3dee-4e64-8559-2a5efe4ceba3/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.102591 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" path="/var/lib/kubelet/pods/c907b3ee-67ab-4d29-a02d-96137c124ff4/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.103895 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" path="/var/lib/kubelet/pods/c9251106-ba81-446b-be7b-51c4ac7f3f81/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.105845 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb846a7f-9be5-4a52-837e-a423f7736e79" path="/var/lib/kubelet/pods/cb846a7f-9be5-4a52-837e-a423f7736e79/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.107466 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" path="/var/lib/kubelet/pods/df6beb49-03ad-47ef-a9c7-3f37baa6d105/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.107755 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-config-data" (OuterVolumeSpecName: "config-data") pod "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" (UID: "875d9bb5-32f8-44c7-aa70-0dfcaf8f9834"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.110001 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" path="/var/lib/kubelet/pods/e3b65efd-d659-4b2d-9cee-0c57b08d029f/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.111593 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fad684df-76f3-4651-81e8-9bb739cfbc4b" path="/var/lib/kubelet/pods/fad684df-76f3-4651-81e8-9bb739cfbc4b/volumes" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.115736 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data" (OuterVolumeSpecName: "config-data") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.118582 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.150708 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.182642 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c81db200-8fff-4a29-952c-4f8609dc8ac8" (UID: "c81db200-8fff-4a29-952c-4f8609dc8ac8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191234 4741 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191267 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191279 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191294 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191314 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191324 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191350 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191362 4741 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c81db200-8fff-4a29-952c-4f8609dc8ac8-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.191380 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hklql\" (UniqueName: \"kubernetes.io/projected/c81db200-8fff-4a29-952c-4f8609dc8ac8-kube-api-access-hklql\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.295308 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-74d7c6f55-9trnx" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.295347 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-74d7c6f55-9trnx" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.357747 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-68snt_865e963c-87bd-45aa-a4a7-95ae24dd0058/ovn-controller/0.log" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.357920 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393323 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run\") pod \"865e963c-87bd-45aa-a4a7-95ae24dd0058\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393483 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/865e963c-87bd-45aa-a4a7-95ae24dd0058-scripts\") pod \"865e963c-87bd-45aa-a4a7-95ae24dd0058\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393522 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-combined-ca-bundle\") pod \"865e963c-87bd-45aa-a4a7-95ae24dd0058\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393550 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-log-ovn\") pod \"865e963c-87bd-45aa-a4a7-95ae24dd0058\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393614 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-ovn-controller-tls-certs\") pod \"865e963c-87bd-45aa-a4a7-95ae24dd0058\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393663 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kdjp\" (UniqueName: \"kubernetes.io/projected/865e963c-87bd-45aa-a4a7-95ae24dd0058-kube-api-access-8kdjp\") pod \"865e963c-87bd-45aa-a4a7-95ae24dd0058\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393688 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run-ovn\") pod \"865e963c-87bd-45aa-a4a7-95ae24dd0058\" (UID: \"865e963c-87bd-45aa-a4a7-95ae24dd0058\") " Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393710 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run" (OuterVolumeSpecName: "var-run") pod "865e963c-87bd-45aa-a4a7-95ae24dd0058" (UID: "865e963c-87bd-45aa-a4a7-95ae24dd0058"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.393753 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "865e963c-87bd-45aa-a4a7-95ae24dd0058" (UID: "865e963c-87bd-45aa-a4a7-95ae24dd0058"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.394022 4741 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.394045 4741 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.394094 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "865e963c-87bd-45aa-a4a7-95ae24dd0058" (UID: "865e963c-87bd-45aa-a4a7-95ae24dd0058"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.395225 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/865e963c-87bd-45aa-a4a7-95ae24dd0058-scripts" (OuterVolumeSpecName: "scripts") pod "865e963c-87bd-45aa-a4a7-95ae24dd0058" (UID: "865e963c-87bd-45aa-a4a7-95ae24dd0058"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.410957 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/865e963c-87bd-45aa-a4a7-95ae24dd0058-kube-api-access-8kdjp" (OuterVolumeSpecName: "kube-api-access-8kdjp") pod "865e963c-87bd-45aa-a4a7-95ae24dd0058" (UID: "865e963c-87bd-45aa-a4a7-95ae24dd0058"). InnerVolumeSpecName "kube-api-access-8kdjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.443512 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "865e963c-87bd-45aa-a4a7-95ae24dd0058" (UID: "865e963c-87bd-45aa-a4a7-95ae24dd0058"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.462155 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "865e963c-87bd-45aa-a4a7-95ae24dd0058" (UID: "865e963c-87bd-45aa-a4a7-95ae24dd0058"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.495301 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/865e963c-87bd-45aa-a4a7-95ae24dd0058-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.495347 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.495363 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e963c-87bd-45aa-a4a7-95ae24dd0058-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.495405 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kdjp\" (UniqueName: \"kubernetes.io/projected/865e963c-87bd-45aa-a4a7-95ae24dd0058-kube-api-access-8kdjp\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.495418 4741 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/865e963c-87bd-45aa-a4a7-95ae24dd0058-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.630214 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-68snt_865e963c-87bd-45aa-a4a7-95ae24dd0058/ovn-controller/0.log" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.630470 4741 generic.go:334] "Generic (PLEG): container finished" podID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerID="1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0" exitCode=137 Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.630522 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt" event={"ID":"865e963c-87bd-45aa-a4a7-95ae24dd0058","Type":"ContainerDied","Data":"1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0"} Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.630549 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-68snt" event={"ID":"865e963c-87bd-45aa-a4a7-95ae24dd0058","Type":"ContainerDied","Data":"7461f39eaefa65fec16c5f7cc9b53d37b849b5d431bfffec7926feecb7ea7a52"} Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.630566 4741 scope.go:117] "RemoveContainer" containerID="1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.630572 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-68snt" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.633376 4741 generic.go:334] "Generic (PLEG): container finished" podID="c81db200-8fff-4a29-952c-4f8609dc8ac8" containerID="18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34" exitCode=0 Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.633523 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677b8f6f88-c7kf7" event={"ID":"c81db200-8fff-4a29-952c-4f8609dc8ac8","Type":"ContainerDied","Data":"18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34"} Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.633544 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677b8f6f88-c7kf7" event={"ID":"c81db200-8fff-4a29-952c-4f8609dc8ac8","Type":"ContainerDied","Data":"c64598fbf2851b5a62163b3e44ad7be97327792b5cc7e7ee844a1bc4dd89c332"} Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.633584 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677b8f6f88-c7kf7" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.645449 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"875d9bb5-32f8-44c7-aa70-0dfcaf8f9834","Type":"ContainerDied","Data":"9ecb21a53a3b4d17f2a69f3f44646b8f16e0c8ee781bf24d844659ca14f7294a"} Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.645522 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.691180 4741 scope.go:117] "RemoveContainer" containerID="1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0" Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.691638 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0\": container with ID starting with 1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0 not found: ID does not exist" containerID="1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.691687 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0"} err="failed to get container status \"1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0\": rpc error: code = NotFound desc = could not find container \"1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0\": container with ID starting with 1f1ee98d8bda00d841884dd62642fbb797cd25ae5eff9f449b3c1462783591c0 not found: ID does not exist" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.691719 4741 scope.go:117] "RemoveContainer" containerID="18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.699033 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-677b8f6f88-c7kf7"] Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.699544 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-677b8f6f88-c7kf7"] Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.708602 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.715880 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.721186 4741 scope.go:117] "RemoveContainer" containerID="18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34" Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.721682 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34\": container with ID starting with 18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34 not found: ID does not exist" containerID="18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.721744 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34"} err="failed to get container status \"18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34\": rpc error: code = NotFound desc = could not find container \"18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34\": container with ID starting with 18a6904d0394108065b864d4aaaa78b078a96d5f0df674165df6d6394d270f34 not found: ID does not exist" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.721776 4741 scope.go:117] "RemoveContainer" containerID="167d2c7f919eb9ab445a2017a6763de4f7c20dd26fb277979f6106c71a0731b5" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.723762 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-68snt"] Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.729945 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-68snt"] Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.742294 4741 scope.go:117] "RemoveContainer" containerID="113041bee656cb97e263bb0fbf9286ccb77bdf43fa28dcf029dc790fd12a60d8" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.774673 4741 scope.go:117] "RemoveContainer" containerID="3d41a9cc13ba07eab14e449222ca1eba4bc2f913c9d707b7f2f9fe2953779503" Sep 29 19:31:13 crc kubenswrapper[4741]: I0929 19:31:13.805332 4741 scope.go:117] "RemoveContainer" containerID="64aaaacdbcb9f52035038a67edb1624996ab0fb313a96a375112066aa4d0c3a6" Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.928521 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.928775 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.928983 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.929014 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.930262 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.931692 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.935727 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:13 crc kubenswrapper[4741]: E0929 19:31:13.935767 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:31:14 crc kubenswrapper[4741]: I0929 19:31:14.222383 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-659f5d9f44-j8zgh" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 19:31:14 crc kubenswrapper[4741]: I0929 19:31:14.222431 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-659f5d9f44-j8zgh" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 19:31:15 crc kubenswrapper[4741]: I0929 19:31:15.099249 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" path="/var/lib/kubelet/pods/865e963c-87bd-45aa-a4a7-95ae24dd0058/volumes" Sep 29 19:31:15 crc kubenswrapper[4741]: I0929 19:31:15.100273 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" path="/var/lib/kubelet/pods/875d9bb5-32f8-44c7-aa70-0dfcaf8f9834/volumes" Sep 29 19:31:15 crc kubenswrapper[4741]: I0929 19:31:15.101094 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c81db200-8fff-4a29-952c-4f8609dc8ac8" path="/var/lib/kubelet/pods/c81db200-8fff-4a29-952c-4f8609dc8ac8/volumes" Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.928148 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.929202 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.929736 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.929830 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.929905 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.931234 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.932736 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:18 crc kubenswrapper[4741]: E0929 19:31:18.932778 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:31:19 crc kubenswrapper[4741]: I0929 19:31:19.146852 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5dd77b8d5c-hw4v7" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.150:9696/\": dial tcp 10.217.0.150:9696: connect: connection refused" Sep 29 19:31:19 crc kubenswrapper[4741]: E0929 19:31:19.999484 4741 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Sep 29 19:31:20 crc kubenswrapper[4741]: E0929 19:31:19.999558 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:35.999540961 +0000 UTC m=+1337.647330293 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-config" not found Sep 29 19:31:20 crc kubenswrapper[4741]: E0929 19:31:19.999893 4741 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Sep 29 19:31:20 crc kubenswrapper[4741]: E0929 19:31:19.999950 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config podName:760672f8-7a2f-4ddb-8ebd-3664670a4adc nodeName:}" failed. No retries permitted until 2025-09-29 19:31:35.999939534 +0000 UTC m=+1337.647728866 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config") pod "neutron-5dd77b8d5c-hw4v7" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc") : secret "neutron-httpd-config" not found Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.928600 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.930023 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.930020 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.931247 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.931354 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.931600 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.934795 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:23 crc kubenswrapper[4741]: E0929 19:31:23.934826 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.774972 4741 generic.go:334] "Generic (PLEG): container finished" podID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerID="3d283ad6e8c834fde894dcf133b306ef1602160587b86f10edad93244d544e8f" exitCode=0 Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.775052 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd77b8d5c-hw4v7" event={"ID":"760672f8-7a2f-4ddb-8ebd-3664670a4adc","Type":"ContainerDied","Data":"3d283ad6e8c834fde894dcf133b306ef1602160587b86f10edad93244d544e8f"} Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.842444 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.912532 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-ovndb-tls-certs\") pod \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.912581 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-combined-ca-bundle\") pod \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.912615 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config\") pod \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.912635 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-public-tls-certs\") pod \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.912673 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config\") pod \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.912732 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-internal-tls-certs\") pod \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.912768 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76fvb\" (UniqueName: \"kubernetes.io/projected/760672f8-7a2f-4ddb-8ebd-3664670a4adc-kube-api-access-76fvb\") pod \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\" (UID: \"760672f8-7a2f-4ddb-8ebd-3664670a4adc\") " Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.922646 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/760672f8-7a2f-4ddb-8ebd-3664670a4adc-kube-api-access-76fvb" (OuterVolumeSpecName: "kube-api-access-76fvb") pod "760672f8-7a2f-4ddb-8ebd-3664670a4adc" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc"). InnerVolumeSpecName "kube-api-access-76fvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.922725 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "760672f8-7a2f-4ddb-8ebd-3664670a4adc" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.948315 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config" (OuterVolumeSpecName: "config") pod "760672f8-7a2f-4ddb-8ebd-3664670a4adc" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.949446 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "760672f8-7a2f-4ddb-8ebd-3664670a4adc" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.953339 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "760672f8-7a2f-4ddb-8ebd-3664670a4adc" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.954156 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "760672f8-7a2f-4ddb-8ebd-3664670a4adc" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:25 crc kubenswrapper[4741]: I0929 19:31:25.984593 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "760672f8-7a2f-4ddb-8ebd-3664670a4adc" (UID: "760672f8-7a2f-4ddb-8ebd-3664670a4adc"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.014565 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.014637 4741 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.014653 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76fvb\" (UniqueName: \"kubernetes.io/projected/760672f8-7a2f-4ddb-8ebd-3664670a4adc-kube-api-access-76fvb\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.014665 4741 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.014677 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.014688 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.014699 4741 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/760672f8-7a2f-4ddb-8ebd-3664670a4adc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.788961 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd77b8d5c-hw4v7" event={"ID":"760672f8-7a2f-4ddb-8ebd-3664670a4adc","Type":"ContainerDied","Data":"b86fa10f7eec29753146890f1e29da4af0d830a57cc9f2a854021a7e52abd618"} Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.789019 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd77b8d5c-hw4v7" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.789298 4741 scope.go:117] "RemoveContainer" containerID="1700c1e978f1aa2ccdb75b0383b4cb4e5f89351b953d72fcb8ea36e16ead7085" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.816327 4741 scope.go:117] "RemoveContainer" containerID="3d283ad6e8c834fde894dcf133b306ef1602160587b86f10edad93244d544e8f" Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.824277 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dd77b8d5c-hw4v7"] Sep 29 19:31:26 crc kubenswrapper[4741]: I0929 19:31:26.837938 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5dd77b8d5c-hw4v7"] Sep 29 19:31:27 crc kubenswrapper[4741]: I0929 19:31:27.096489 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" path="/var/lib/kubelet/pods/760672f8-7a2f-4ddb-8ebd-3664670a4adc/volumes" Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.929359 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.930075 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.930538 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.930614 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.930745 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.931979 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.933802 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:28 crc kubenswrapper[4741]: E0929 19:31:28.933882 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.928272 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.929477 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.929714 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.930499 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.930559 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.931790 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.933325 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Sep 29 19:31:33 crc kubenswrapper[4741]: E0929 19:31:33.933378 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-q65b6" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:31:34 crc kubenswrapper[4741]: I0929 19:31:34.870372 4741 generic.go:334] "Generic (PLEG): container finished" podID="95b48711-844f-4e95-9f01-36bec6b9995a" containerID="e8277ed6dfec4fbba8595886bec52f0c929d6c4a577686a175df2b460882c90f" exitCode=137 Sep 29 19:31:34 crc kubenswrapper[4741]: I0929 19:31:34.870420 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"e8277ed6dfec4fbba8595886bec52f0c929d6c4a577686a175df2b460882c90f"} Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.083938 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.154782 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-cache\") pod \"95b48711-844f-4e95-9f01-36bec6b9995a\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.154853 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-lock\") pod \"95b48711-844f-4e95-9f01-36bec6b9995a\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.154946 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"95b48711-844f-4e95-9f01-36bec6b9995a\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.155015 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") pod \"95b48711-844f-4e95-9f01-36bec6b9995a\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.155069 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d62jq\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-kube-api-access-d62jq\") pod \"95b48711-844f-4e95-9f01-36bec6b9995a\" (UID: \"95b48711-844f-4e95-9f01-36bec6b9995a\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.155677 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-lock" (OuterVolumeSpecName: "lock") pod "95b48711-844f-4e95-9f01-36bec6b9995a" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.156097 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-cache" (OuterVolumeSpecName: "cache") pod "95b48711-844f-4e95-9f01-36bec6b9995a" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.160102 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "95b48711-844f-4e95-9f01-36bec6b9995a" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.160165 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-kube-api-access-d62jq" (OuterVolumeSpecName: "kube-api-access-d62jq") pod "95b48711-844f-4e95-9f01-36bec6b9995a" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a"). InnerVolumeSpecName "kube-api-access-d62jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.167600 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "swift") pod "95b48711-844f-4e95-9f01-36bec6b9995a" (UID: "95b48711-844f-4e95-9f01-36bec6b9995a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.256958 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.257007 4741 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.257028 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d62jq\" (UniqueName: \"kubernetes.io/projected/95b48711-844f-4e95-9f01-36bec6b9995a-kube-api-access-d62jq\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.257047 4741 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-cache\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.257063 4741 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/95b48711-844f-4e95-9f01-36bec6b9995a-lock\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.277291 4741 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.358401 4741 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.578970 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q65b6_1e20498c-3a09-44e9-a4b1-c2dad8437611/ovs-vswitchd/0.log" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.579728 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.662911 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e20498c-3a09-44e9-a4b1-c2dad8437611-scripts\") pod \"1e20498c-3a09-44e9-a4b1-c2dad8437611\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.662969 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-lib\") pod \"1e20498c-3a09-44e9-a4b1-c2dad8437611\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.663000 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlcmc\" (UniqueName: \"kubernetes.io/projected/1e20498c-3a09-44e9-a4b1-c2dad8437611-kube-api-access-qlcmc\") pod \"1e20498c-3a09-44e9-a4b1-c2dad8437611\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.663033 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-log\") pod \"1e20498c-3a09-44e9-a4b1-c2dad8437611\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.663161 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-etc-ovs\") pod \"1e20498c-3a09-44e9-a4b1-c2dad8437611\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.663259 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-run\") pod \"1e20498c-3a09-44e9-a4b1-c2dad8437611\" (UID: \"1e20498c-3a09-44e9-a4b1-c2dad8437611\") " Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.663640 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-run" (OuterVolumeSpecName: "var-run") pod "1e20498c-3a09-44e9-a4b1-c2dad8437611" (UID: "1e20498c-3a09-44e9-a4b1-c2dad8437611"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.663709 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-log" (OuterVolumeSpecName: "var-log") pod "1e20498c-3a09-44e9-a4b1-c2dad8437611" (UID: "1e20498c-3a09-44e9-a4b1-c2dad8437611"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.663733 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "1e20498c-3a09-44e9-a4b1-c2dad8437611" (UID: "1e20498c-3a09-44e9-a4b1-c2dad8437611"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.664025 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-lib" (OuterVolumeSpecName: "var-lib") pod "1e20498c-3a09-44e9-a4b1-c2dad8437611" (UID: "1e20498c-3a09-44e9-a4b1-c2dad8437611"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.665010 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e20498c-3a09-44e9-a4b1-c2dad8437611-scripts" (OuterVolumeSpecName: "scripts") pod "1e20498c-3a09-44e9-a4b1-c2dad8437611" (UID: "1e20498c-3a09-44e9-a4b1-c2dad8437611"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.667167 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e20498c-3a09-44e9-a4b1-c2dad8437611-kube-api-access-qlcmc" (OuterVolumeSpecName: "kube-api-access-qlcmc") pod "1e20498c-3a09-44e9-a4b1-c2dad8437611" (UID: "1e20498c-3a09-44e9-a4b1-c2dad8437611"). InnerVolumeSpecName "kube-api-access-qlcmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.764805 4741 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-etc-ovs\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.764848 4741 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-run\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.764861 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e20498c-3a09-44e9-a4b1-c2dad8437611-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.764872 4741 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-lib\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.764885 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlcmc\" (UniqueName: \"kubernetes.io/projected/1e20498c-3a09-44e9-a4b1-c2dad8437611-kube-api-access-qlcmc\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.764900 4741 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e20498c-3a09-44e9-a4b1-c2dad8437611-var-log\") on node \"crc\" DevicePath \"\"" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.880690 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q65b6_1e20498c-3a09-44e9-a4b1-c2dad8437611/ovs-vswitchd/0.log" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.881507 4741 generic.go:334] "Generic (PLEG): container finished" podID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" exitCode=137 Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.881612 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q65b6" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.881607 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q65b6" event={"ID":"1e20498c-3a09-44e9-a4b1-c2dad8437611","Type":"ContainerDied","Data":"e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f"} Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.881707 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q65b6" event={"ID":"1e20498c-3a09-44e9-a4b1-c2dad8437611","Type":"ContainerDied","Data":"c62f162f41d0b509b84e2fd845f6fdbce9fe539f3c55c6229bc63980d43622ab"} Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.881750 4741 scope.go:117] "RemoveContainer" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.888922 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"95b48711-844f-4e95-9f01-36bec6b9995a","Type":"ContainerDied","Data":"d1de83713e8780fdcbf82c2f2fcedf1895acdd57776e89f880ce3263918c9aac"} Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.888997 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.922529 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-q65b6"] Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.926909 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-q65b6"] Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.945364 4741 scope.go:117] "RemoveContainer" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.950525 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.956372 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Sep 29 19:31:35 crc kubenswrapper[4741]: I0929 19:31:35.977226 4741 scope.go:117] "RemoveContainer" containerID="891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.014784 4741 scope.go:117] "RemoveContainer" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" Sep 29 19:31:36 crc kubenswrapper[4741]: E0929 19:31:36.015438 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f\": container with ID starting with e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f not found: ID does not exist" containerID="e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.015495 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f"} err="failed to get container status \"e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f\": rpc error: code = NotFound desc = could not find container \"e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f\": container with ID starting with e01c896a398ba12b803fbdf4e45e8a3f8c0e1926687ddd181afc009dfd3a6c6f not found: ID does not exist" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.015522 4741 scope.go:117] "RemoveContainer" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" Sep 29 19:31:36 crc kubenswrapper[4741]: E0929 19:31:36.015964 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14\": container with ID starting with f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 not found: ID does not exist" containerID="f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.015998 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14"} err="failed to get container status \"f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14\": rpc error: code = NotFound desc = could not find container \"f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14\": container with ID starting with f288f2b843453e1dde90e8de385283c56d4fdcb48346a59848773e48f6d81a14 not found: ID does not exist" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.016020 4741 scope.go:117] "RemoveContainer" containerID="891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa" Sep 29 19:31:36 crc kubenswrapper[4741]: E0929 19:31:36.016375 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa\": container with ID starting with 891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa not found: ID does not exist" containerID="891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.016417 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa"} err="failed to get container status \"891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa\": rpc error: code = NotFound desc = could not find container \"891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa\": container with ID starting with 891cef90bad78c8fa8c30981069b05c1683a54067336e2c01e73e08ac7094ffa not found: ID does not exist" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.016432 4741 scope.go:117] "RemoveContainer" containerID="e8277ed6dfec4fbba8595886bec52f0c929d6c4a577686a175df2b460882c90f" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.043652 4741 scope.go:117] "RemoveContainer" containerID="af59f98d013533f7b1f00c005e49813bc7d0e7ab704e34a34db4cec3151d2b45" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.060132 4741 scope.go:117] "RemoveContainer" containerID="a979b22c48d4833060b798e59c6172ae0d0fb020aa7081564218be678d051ef8" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.074499 4741 scope.go:117] "RemoveContainer" containerID="878b7f5b4bcf11108cbc1b12fcff16a252c21bb2290b20babdfe2cc60f5f4278" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.090139 4741 scope.go:117] "RemoveContainer" containerID="a68bac92910f13da6b5bd1c4c6fa29f7fec8ba0ea712f3e3c9da7813f47eafee" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.115931 4741 scope.go:117] "RemoveContainer" containerID="5ace69d8ba764865fefb3a17c320882c81b7b6f3fba4329e49c21e37a73ff7f1" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.135758 4741 scope.go:117] "RemoveContainer" containerID="71e20963cd9fad0ee127b4dda01c84dd767efbefb2fcb80296923f8718f1ddbf" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.150798 4741 scope.go:117] "RemoveContainer" containerID="49334c9dcc9f0ab359baa344e3fe68ec6f0e5271440eec25dd046555f7cc0e20" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.168262 4741 scope.go:117] "RemoveContainer" containerID="8bdeba0cf6a202fe00074a937a1baa94541c1fc21049258f466a2ee65af3a75e" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.186210 4741 scope.go:117] "RemoveContainer" containerID="e650c5ffc6ebc6918dc63202d7b9d6aaff80d4b6db1471fd84f3faebfd8b934f" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.204022 4741 scope.go:117] "RemoveContainer" containerID="0272c3b7736bc9733f76397bbd76a46beab031e046ac627b3f5b05578332ba13" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.219127 4741 scope.go:117] "RemoveContainer" containerID="11c745f1505ad37096a8441c7db5e1f99996efc6719e0f171773f7647d65d6f0" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.251026 4741 scope.go:117] "RemoveContainer" containerID="5c99d27cbf271df86190fc67b6bd3129d7c9ebc403240020a83b932781c09aa8" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.268647 4741 scope.go:117] "RemoveContainer" containerID="bdcc643cf7b0ba52293b398520b054d5a0af7419de83697c00319b14e82a22f5" Sep 29 19:31:36 crc kubenswrapper[4741]: I0929 19:31:36.286946 4741 scope.go:117] "RemoveContainer" containerID="0f18bf31b15ac6485b4321250e71d201fadbe48a440243ff5584feadda173ffc" Sep 29 19:31:37 crc kubenswrapper[4741]: I0929 19:31:37.095051 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" path="/var/lib/kubelet/pods/1e20498c-3a09-44e9-a4b1-c2dad8437611/volumes" Sep 29 19:31:37 crc kubenswrapper[4741]: I0929 19:31:37.096211 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" path="/var/lib/kubelet/pods/95b48711-844f-4e95-9f01-36bec6b9995a/volumes" Sep 29 19:31:39 crc kubenswrapper[4741]: I0929 19:31:39.008305 4741 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod2094977c-9460-4dfd-926d-aac495e9bb73"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod2094977c-9460-4dfd-926d-aac495e9bb73] : Timed out while waiting for systemd to remove kubepods-besteffort-pod2094977c_9460_4dfd_926d_aac495e9bb73.slice" Sep 29 19:31:39 crc kubenswrapper[4741]: E0929 19:31:39.008440 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod2094977c-9460-4dfd-926d-aac495e9bb73] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod2094977c-9460-4dfd-926d-aac495e9bb73] : Timed out while waiting for systemd to remove kubepods-besteffort-pod2094977c_9460_4dfd_926d_aac495e9bb73.slice" pod="openstack/neutron07da-account-delete-vj777" podUID="2094977c-9460-4dfd-926d-aac495e9bb73" Sep 29 19:31:39 crc kubenswrapper[4741]: I0929 19:31:39.928642 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron07da-account-delete-vj777" Sep 29 19:31:39 crc kubenswrapper[4741]: I0929 19:31:39.966358 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron07da-account-delete-vj777"] Sep 29 19:31:39 crc kubenswrapper[4741]: I0929 19:31:39.970314 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron07da-account-delete-vj777"] Sep 29 19:31:41 crc kubenswrapper[4741]: I0929 19:31:41.096366 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2094977c-9460-4dfd-926d-aac495e9bb73" path="/var/lib/kubelet/pods/2094977c-9460-4dfd-926d-aac495e9bb73/volumes" Sep 29 19:32:01 crc kubenswrapper[4741]: I0929 19:32:01.739242 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:32:01 crc kubenswrapper[4741]: I0929 19:32:01.740657 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.590649 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9lb7w"] Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.605214 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerName="mysql-bootstrap" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.605309 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerName="mysql-bootstrap" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.605374 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.605455 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.605530 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="probe" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.605593 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="probe" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.605669 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.605734 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-server" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.605814 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerName="ovn-controller" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.605870 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerName="ovn-controller" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.605935 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-updater" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.605988 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-updater" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606046 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="rsync" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606098 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="rsync" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606149 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606205 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606274 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606352 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606429 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d18b4c1-d423-4840-97ff-c322272c3aa3" containerName="nova-scheduler-scheduler" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606514 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d18b4c1-d423-4840-97ff-c322272c3aa3" containerName="nova-scheduler-scheduler" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606599 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606675 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606757 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606845 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606890 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606914 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-server" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606928 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606937 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606970 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerName="setup-container" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.606979 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerName="setup-container" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.606997 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607007 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607042 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607051 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-api" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607062 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607070 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-api" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607096 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607105 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607128 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607136 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607151 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607160 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607178 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="cinder-scheduler" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607186 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="cinder-scheduler" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607196 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13785b5b-f629-47ae-8251-331d44b40254" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607206 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="13785b5b-f629-47ae-8251-331d44b40254" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607224 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerName="setup-container" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607235 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerName="setup-container" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607260 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="swift-recon-cron" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607270 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="swift-recon-cron" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607290 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-metadata" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607299 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-metadata" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607311 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607319 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607336 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-notification-agent" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607345 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-notification-agent" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607373 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" containerName="kube-state-metrics" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607447 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" containerName="kube-state-metrics" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607457 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607466 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-server" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607494 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c5456d-7a59-4991-ad3d-58b04db78b24" containerName="nova-cell1-conductor-conductor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607503 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c5456d-7a59-4991-ad3d-58b04db78b24" containerName="nova-cell1-conductor-conductor" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607520 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607528 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607562 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="openstack-network-exporter" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607571 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="openstack-network-exporter" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607586 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-reaper" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607594 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-reaper" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607624 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607632 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607650 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607663 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-api" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607685 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-central-agent" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607694 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-central-agent" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607712 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607721 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607739 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607747 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607770 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81db200-8fff-4a29-952c-4f8609dc8ac8" containerName="keystone-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607778 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81db200-8fff-4a29-952c-4f8609dc8ac8" containerName="keystone-api" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607804 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607812 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607846 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607854 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607884 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="sg-core" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607893 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="sg-core" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607917 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerName="galera" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607925 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerName="galera" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607952 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" containerName="memcached" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607961 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" containerName="memcached" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607972 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb846a7f-9be5-4a52-837e-a423f7736e79" containerName="nova-cell0-conductor-conductor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.607981 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb846a7f-9be5-4a52-837e-a423f7736e79" containerName="nova-cell0-conductor-conductor" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.607992 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608000 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608016 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="ovn-northd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608025 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="ovn-northd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608041 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="proxy-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608049 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="proxy-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608074 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server-init" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608549 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server-init" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608590 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608597 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608623 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-updater" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608629 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-updater" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608640 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608647 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608659 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983d2897-24d2-462c-b9f9-427639d2b8f9" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608665 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="983d2897-24d2-462c-b9f9-427639d2b8f9" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608680 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608686 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608704 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608749 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608761 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608770 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608781 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerName="rabbitmq" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608788 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerName="rabbitmq" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608800 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608807 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-server" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608829 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608836 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608852 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608858 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608874 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608880 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608889 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerName="rabbitmq" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608895 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerName="rabbitmq" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.608907 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-expirer" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.608913 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-expirer" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616067 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c5456d-7a59-4991-ad3d-58b04db78b24" containerName="nova-cell1-conductor-conductor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616118 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616135 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616154 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616166 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="df6beb49-03ad-47ef-a9c7-3f37baa6d105" containerName="rabbitmq" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616196 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616213 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616238 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616262 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="760672f8-7a2f-4ddb-8ebd-3664670a4adc" containerName="neutron-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616281 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="rsync" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616305 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="983d2897-24d2-462c-b9f9-427639d2b8f9" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616321 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c81db200-8fff-4a29-952c-4f8609dc8ac8" containerName="keystone-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616347 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-central-agent" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616366 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="ovn-northd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616382 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-expirer" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616429 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0c02dc-69a9-4e60-b179-0e23842d10a4" containerName="rabbitmq" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616439 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616463 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovs-vswitchd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616471 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-updater" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616490 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616512 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616521 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616563 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616578 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="proxy-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616608 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d18b4c1-d423-4840-97ff-c322272c3aa3" containerName="nova-scheduler-scheduler" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616622 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="cinder-scheduler" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616634 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c39883d-b9e0-4b1f-a7b0-8d29c04d066a" containerName="kube-state-metrics" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616644 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8cf4ba-5168-48b5-8424-3e3e21bbdae3" containerName="memcached" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616679 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2efb9fff-2e44-4257-bbe6-e5dccedcdcbe" containerName="proxy-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616690 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efd2b95-d8dd-42bf-ab6a-2f7ef3c83d62" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616714 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-updater" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616731 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-metadata" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616742 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e20498c-3a09-44e9-a4b1-c2dad8437611" containerName="ovsdb-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616763 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0a22c3-d524-422d-8692-14d7a16a418f" containerName="nova-metadata-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616787 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="account-reaper" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616806 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2be2676f-d882-424e-8a49-d8bffc23773e" containerName="placement-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616821 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="58cfa8a3-5067-40ee-8e23-803a52d6b1b1" containerName="barbican-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616839 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ade550-0eaf-4518-8696-14daf3f034d4" containerName="glance-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616862 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-server" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616881 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616901 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="object-replicator" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616926 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616943 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c907b3ee-67ab-4d29-a02d-96137c124ff4" containerName="probe" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616959 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="13785b5b-f629-47ae-8251-331d44b40254" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616972 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.616995 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617012 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api-log" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617339 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="swift-recon-cron" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617357 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b65efd-d659-4b2d-9cee-0c57b08d029f" containerName="cinder-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617375 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="fad684df-76f3-4651-81e8-9bb739cfbc4b" containerName="galera" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617385 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb846a7f-9be5-4a52-837e-a423f7736e79" containerName="nova-cell0-conductor-conductor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617425 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="865e963c-87bd-45aa-a4a7-95ae24dd0058" containerName="ovn-controller" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617475 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b48711-844f-4e95-9f01-36bec6b9995a" containerName="container-auditor" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617493 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="be00d3f7-3dee-4e64-8559-2a5efe4ceba3" containerName="nova-api-api" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617511 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="sg-core" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617521 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="13785b5b-f629-47ae-8251-331d44b40254" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.617542 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d4bbc33-c647-4eda-bdc6-89f80678ae8d" containerName="openstack-network-exporter" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.618003 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e3a43d-5ff8-491d-916c-3c0b9e31f223" containerName="barbican-worker" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.618042 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c715e1b9-b2a4-44a0-9ae8-b1a854c1a19d" containerName="barbican-keystone-listener" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.618056 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="875d9bb5-32f8-44c7-aa70-0dfcaf8f9834" containerName="ceilometer-notification-agent" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.618083 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9251106-ba81-446b-be7b-51c4ac7f3f81" containerName="glance-httpd" Sep 29 19:32:22 crc kubenswrapper[4741]: E0929 19:32:22.618521 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13785b5b-f629-47ae-8251-331d44b40254" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.618540 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="13785b5b-f629-47ae-8251-331d44b40254" containerName="mariadb-account-delete" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.620023 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lb7w"] Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.620135 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.660331 4741 scope.go:117] "RemoveContainer" containerID="7c71b4d390feacc6365d3ebfe09f6e223b707ee503a77ad41f8f50b865956428" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.663349 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-catalog-content\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.663382 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9hdt\" (UniqueName: \"kubernetes.io/projected/0df65740-0927-4b0d-9781-0011bf3d566d-kube-api-access-x9hdt\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.663464 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-utilities\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.764992 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-catalog-content\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.765060 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9hdt\" (UniqueName: \"kubernetes.io/projected/0df65740-0927-4b0d-9781-0011bf3d566d-kube-api-access-x9hdt\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.765149 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-utilities\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.765714 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-utilities\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.765987 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-catalog-content\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.796613 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9hdt\" (UniqueName: \"kubernetes.io/projected/0df65740-0927-4b0d-9781-0011bf3d566d-kube-api-access-x9hdt\") pod \"redhat-marketplace-9lb7w\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:22 crc kubenswrapper[4741]: I0929 19:32:22.942139 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:23 crc kubenswrapper[4741]: I0929 19:32:23.386660 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lb7w"] Sep 29 19:32:24 crc kubenswrapper[4741]: I0929 19:32:24.365173 4741 generic.go:334] "Generic (PLEG): container finished" podID="0df65740-0927-4b0d-9781-0011bf3d566d" containerID="1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180" exitCode=0 Sep 29 19:32:24 crc kubenswrapper[4741]: I0929 19:32:24.365271 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lb7w" event={"ID":"0df65740-0927-4b0d-9781-0011bf3d566d","Type":"ContainerDied","Data":"1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180"} Sep 29 19:32:24 crc kubenswrapper[4741]: I0929 19:32:24.365762 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lb7w" event={"ID":"0df65740-0927-4b0d-9781-0011bf3d566d","Type":"ContainerStarted","Data":"1bb746efc4651446ca9a495508cd3329b0d945221b05ea69cdfd8d68d98f693f"} Sep 29 19:32:24 crc kubenswrapper[4741]: I0929 19:32:24.367983 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 19:32:26 crc kubenswrapper[4741]: I0929 19:32:26.382593 4741 generic.go:334] "Generic (PLEG): container finished" podID="0df65740-0927-4b0d-9781-0011bf3d566d" containerID="a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc" exitCode=0 Sep 29 19:32:26 crc kubenswrapper[4741]: I0929 19:32:26.382634 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lb7w" event={"ID":"0df65740-0927-4b0d-9781-0011bf3d566d","Type":"ContainerDied","Data":"a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc"} Sep 29 19:32:27 crc kubenswrapper[4741]: I0929 19:32:27.391765 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lb7w" event={"ID":"0df65740-0927-4b0d-9781-0011bf3d566d","Type":"ContainerStarted","Data":"3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b"} Sep 29 19:32:27 crc kubenswrapper[4741]: I0929 19:32:27.409199 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9lb7w" podStartSLOduration=2.8614066339999997 podStartE2EDuration="5.409181782s" podCreationTimestamp="2025-09-29 19:32:22 +0000 UTC" firstStartedPulling="2025-09-29 19:32:24.367650771 +0000 UTC m=+1386.015440113" lastFinishedPulling="2025-09-29 19:32:26.915425929 +0000 UTC m=+1388.563215261" observedRunningTime="2025-09-29 19:32:27.406424515 +0000 UTC m=+1389.054213867" watchObservedRunningTime="2025-09-29 19:32:27.409181782 +0000 UTC m=+1389.056971114" Sep 29 19:32:31 crc kubenswrapper[4741]: I0929 19:32:31.740241 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:32:31 crc kubenswrapper[4741]: I0929 19:32:31.740894 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:32:32 crc kubenswrapper[4741]: I0929 19:32:32.942437 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:32 crc kubenswrapper[4741]: I0929 19:32:32.942496 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:33 crc kubenswrapper[4741]: I0929 19:32:33.012012 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:33 crc kubenswrapper[4741]: I0929 19:32:33.497210 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:33 crc kubenswrapper[4741]: I0929 19:32:33.545963 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lb7w"] Sep 29 19:32:35 crc kubenswrapper[4741]: I0929 19:32:35.460851 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9lb7w" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="registry-server" containerID="cri-o://3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b" gracePeriod=2 Sep 29 19:32:35 crc kubenswrapper[4741]: I0929 19:32:35.924590 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.052149 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-utilities\") pod \"0df65740-0927-4b0d-9781-0011bf3d566d\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.052293 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9hdt\" (UniqueName: \"kubernetes.io/projected/0df65740-0927-4b0d-9781-0011bf3d566d-kube-api-access-x9hdt\") pod \"0df65740-0927-4b0d-9781-0011bf3d566d\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.052498 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-catalog-content\") pod \"0df65740-0927-4b0d-9781-0011bf3d566d\" (UID: \"0df65740-0927-4b0d-9781-0011bf3d566d\") " Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.053840 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-utilities" (OuterVolumeSpecName: "utilities") pod "0df65740-0927-4b0d-9781-0011bf3d566d" (UID: "0df65740-0927-4b0d-9781-0011bf3d566d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.064623 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0df65740-0927-4b0d-9781-0011bf3d566d-kube-api-access-x9hdt" (OuterVolumeSpecName: "kube-api-access-x9hdt") pod "0df65740-0927-4b0d-9781-0011bf3d566d" (UID: "0df65740-0927-4b0d-9781-0011bf3d566d"). InnerVolumeSpecName "kube-api-access-x9hdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.071042 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0df65740-0927-4b0d-9781-0011bf3d566d" (UID: "0df65740-0927-4b0d-9781-0011bf3d566d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.153888 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.153930 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0df65740-0927-4b0d-9781-0011bf3d566d-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.153943 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9hdt\" (UniqueName: \"kubernetes.io/projected/0df65740-0927-4b0d-9781-0011bf3d566d-kube-api-access-x9hdt\") on node \"crc\" DevicePath \"\"" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.471872 4741 generic.go:334] "Generic (PLEG): container finished" podID="0df65740-0927-4b0d-9781-0011bf3d566d" containerID="3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b" exitCode=0 Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.472774 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lb7w" event={"ID":"0df65740-0927-4b0d-9781-0011bf3d566d","Type":"ContainerDied","Data":"3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b"} Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.472885 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lb7w" event={"ID":"0df65740-0927-4b0d-9781-0011bf3d566d","Type":"ContainerDied","Data":"1bb746efc4651446ca9a495508cd3329b0d945221b05ea69cdfd8d68d98f693f"} Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.472958 4741 scope.go:117] "RemoveContainer" containerID="3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.473133 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lb7w" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.498605 4741 scope.go:117] "RemoveContainer" containerID="a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.523847 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lb7w"] Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.535672 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lb7w"] Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.575131 4741 scope.go:117] "RemoveContainer" containerID="1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.610591 4741 scope.go:117] "RemoveContainer" containerID="3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b" Sep 29 19:32:36 crc kubenswrapper[4741]: E0929 19:32:36.621649 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b\": container with ID starting with 3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b not found: ID does not exist" containerID="3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.621702 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b"} err="failed to get container status \"3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b\": rpc error: code = NotFound desc = could not find container \"3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b\": container with ID starting with 3edc3704b6946dec1054a8971a845004c1a0f6ad2b4b9fb6eb4cd43f7b82986b not found: ID does not exist" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.621736 4741 scope.go:117] "RemoveContainer" containerID="a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc" Sep 29 19:32:36 crc kubenswrapper[4741]: E0929 19:32:36.630553 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc\": container with ID starting with a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc not found: ID does not exist" containerID="a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.630613 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc"} err="failed to get container status \"a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc\": rpc error: code = NotFound desc = could not find container \"a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc\": container with ID starting with a7285f93014c2a931d566fb0025a5e7a3e0d4e71909c5369b1a515e9277164cc not found: ID does not exist" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.630643 4741 scope.go:117] "RemoveContainer" containerID="1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180" Sep 29 19:32:36 crc kubenswrapper[4741]: E0929 19:32:36.636556 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180\": container with ID starting with 1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180 not found: ID does not exist" containerID="1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180" Sep 29 19:32:36 crc kubenswrapper[4741]: I0929 19:32:36.636607 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180"} err="failed to get container status \"1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180\": rpc error: code = NotFound desc = could not find container \"1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180\": container with ID starting with 1bdde6676b8b2f204e7eb2c9eabf8147e15c13f63f6d8a0fcf65a8dbc241a180 not found: ID does not exist" Sep 29 19:32:37 crc kubenswrapper[4741]: I0929 19:32:37.095350 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" path="/var/lib/kubelet/pods/0df65740-0927-4b0d-9781-0011bf3d566d/volumes" Sep 29 19:33:01 crc kubenswrapper[4741]: I0929 19:33:01.738604 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:33:01 crc kubenswrapper[4741]: I0929 19:33:01.739157 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:33:01 crc kubenswrapper[4741]: I0929 19:33:01.739197 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:33:01 crc kubenswrapper[4741]: I0929 19:33:01.740016 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a962e46acb3fe13f6d108a85b06c3220bc7a353adc196e7c6597c33dec71e863"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:33:01 crc kubenswrapper[4741]: I0929 19:33:01.740079 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://a962e46acb3fe13f6d108a85b06c3220bc7a353adc196e7c6597c33dec71e863" gracePeriod=600 Sep 29 19:33:02 crc kubenswrapper[4741]: I0929 19:33:02.739680 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="a962e46acb3fe13f6d108a85b06c3220bc7a353adc196e7c6597c33dec71e863" exitCode=0 Sep 29 19:33:02 crc kubenswrapper[4741]: I0929 19:33:02.739805 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"a962e46acb3fe13f6d108a85b06c3220bc7a353adc196e7c6597c33dec71e863"} Sep 29 19:33:02 crc kubenswrapper[4741]: I0929 19:33:02.740214 4741 scope.go:117] "RemoveContainer" containerID="f1bdfb84a67cbc508e610930001920c7a19e715be8c9f4e424f1f5b91fd396f8" Sep 29 19:33:02 crc kubenswrapper[4741]: I0929 19:33:02.740074 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8"} Sep 29 19:33:11 crc kubenswrapper[4741]: I0929 19:33:11.962875 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mtgjf"] Sep 29 19:33:11 crc kubenswrapper[4741]: E0929 19:33:11.963562 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="extract-utilities" Sep 29 19:33:11 crc kubenswrapper[4741]: I0929 19:33:11.963578 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="extract-utilities" Sep 29 19:33:11 crc kubenswrapper[4741]: E0929 19:33:11.963610 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="registry-server" Sep 29 19:33:11 crc kubenswrapper[4741]: I0929 19:33:11.963618 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="registry-server" Sep 29 19:33:11 crc kubenswrapper[4741]: E0929 19:33:11.963640 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="extract-content" Sep 29 19:33:11 crc kubenswrapper[4741]: I0929 19:33:11.963647 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="extract-content" Sep 29 19:33:11 crc kubenswrapper[4741]: I0929 19:33:11.963819 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0df65740-0927-4b0d-9781-0011bf3d566d" containerName="registry-server" Sep 29 19:33:11 crc kubenswrapper[4741]: I0929 19:33:11.965039 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:11 crc kubenswrapper[4741]: I0929 19:33:11.974765 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtgjf"] Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.104128 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-utilities\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.104341 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-catalog-content\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.104415 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j656r\" (UniqueName: \"kubernetes.io/projected/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-kube-api-access-j656r\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.205374 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-utilities\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.205591 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-catalog-content\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.205648 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j656r\" (UniqueName: \"kubernetes.io/projected/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-kube-api-access-j656r\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.205933 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-utilities\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.206235 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-catalog-content\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.230770 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j656r\" (UniqueName: \"kubernetes.io/projected/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-kube-api-access-j656r\") pod \"community-operators-mtgjf\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.300728 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.828248 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtgjf"] Sep 29 19:33:12 crc kubenswrapper[4741]: I0929 19:33:12.836864 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtgjf" event={"ID":"ccf6e879-2dee-4251-9752-8c7c1fcdeda0","Type":"ContainerStarted","Data":"0f0a8031e331c55605b990e1a44a93c2f1357ef877ea43866f1d47c47395ea37"} Sep 29 19:33:13 crc kubenswrapper[4741]: I0929 19:33:13.847748 4741 generic.go:334] "Generic (PLEG): container finished" podID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerID="60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a" exitCode=0 Sep 29 19:33:13 crc kubenswrapper[4741]: I0929 19:33:13.847973 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtgjf" event={"ID":"ccf6e879-2dee-4251-9752-8c7c1fcdeda0","Type":"ContainerDied","Data":"60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a"} Sep 29 19:33:15 crc kubenswrapper[4741]: I0929 19:33:15.865635 4741 generic.go:334] "Generic (PLEG): container finished" podID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerID="29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f" exitCode=0 Sep 29 19:33:15 crc kubenswrapper[4741]: I0929 19:33:15.865719 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtgjf" event={"ID":"ccf6e879-2dee-4251-9752-8c7c1fcdeda0","Type":"ContainerDied","Data":"29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f"} Sep 29 19:33:16 crc kubenswrapper[4741]: I0929 19:33:16.879428 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtgjf" event={"ID":"ccf6e879-2dee-4251-9752-8c7c1fcdeda0","Type":"ContainerStarted","Data":"60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6"} Sep 29 19:33:16 crc kubenswrapper[4741]: I0929 19:33:16.898591 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mtgjf" podStartSLOduration=3.446018381 podStartE2EDuration="5.898568241s" podCreationTimestamp="2025-09-29 19:33:11 +0000 UTC" firstStartedPulling="2025-09-29 19:33:13.850698751 +0000 UTC m=+1435.498488083" lastFinishedPulling="2025-09-29 19:33:16.303248611 +0000 UTC m=+1437.951037943" observedRunningTime="2025-09-29 19:33:16.895935908 +0000 UTC m=+1438.543725240" watchObservedRunningTime="2025-09-29 19:33:16.898568241 +0000 UTC m=+1438.546357573" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.300817 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.301524 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.357745 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.780508 4741 scope.go:117] "RemoveContainer" containerID="eb44b7cca7f1c498e95932d205c676cab41a427d9a4d380e41e4ec0ae2e00bb9" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.802700 4741 scope.go:117] "RemoveContainer" containerID="5cecc8c2ec11165c95aa8393dacbd5b4792e48d4a72193d8c53722290b713ee7" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.833512 4741 scope.go:117] "RemoveContainer" containerID="f996a2898035c79af3529c1361930f4c58e64aae7f679f404e92d59c0e8200c0" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.858769 4741 scope.go:117] "RemoveContainer" containerID="acd32fd79bc7d0c38d24d970d07163f84e2a2deb7951623bb290c323bb8d913a" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.919221 4741 scope.go:117] "RemoveContainer" containerID="ba8ba5bcbd0b0037bd0c1077997c3f1cb5353af5b057ed27806373a4e5d32b5a" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.962863 4741 scope.go:117] "RemoveContainer" containerID="80996d6bbd486a08fd7d5acf00412689ada3eb81d2f0bd189a0667e966accbcc" Sep 29 19:33:22 crc kubenswrapper[4741]: I0929 19:33:22.998313 4741 scope.go:117] "RemoveContainer" containerID="64a5905d0be5104788453fc11b7eaaf36923a36ee162da53e9e8481b08c87205" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.006910 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.025017 4741 scope.go:117] "RemoveContainer" containerID="682294e2a19501673b2a62943ea271a293eba88bea8c3d9e3f37a7422b35cdf7" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.048377 4741 scope.go:117] "RemoveContainer" containerID="d131ad911bbb73579247dd956ee8718242e5a384d4d827fd0a111dfc506babf9" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.050540 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtgjf"] Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.066795 4741 scope.go:117] "RemoveContainer" containerID="f24cdfed73077cdb3c2d36e877e66cfcc8b22dcefabe794c04d88c2e3b0f8099" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.082922 4741 scope.go:117] "RemoveContainer" containerID="da0da8b5d3e104bad06c23cfff349af0ee1a80482a6195e6ac53bebbeaf752f2" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.100047 4741 scope.go:117] "RemoveContainer" containerID="1176f4d9203ffd34279ab2077789bd46d8d3ccd54e327e9be6c17203056b22cf" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.122955 4741 scope.go:117] "RemoveContainer" containerID="c840b9d34193b3248a4a42d0dfa2c7dcbcf07b30ff0be4e77faf1f7aa259dda9" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.152564 4741 scope.go:117] "RemoveContainer" containerID="59ff3431a84ba005bd2b5c3d863965362c61aebed601f9c5ea3a2b72e47dc82b" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.190250 4741 scope.go:117] "RemoveContainer" containerID="d21304531a9d12f50d20982ea29f8c74f91db31217925de0d1242446471f527d" Sep 29 19:33:23 crc kubenswrapper[4741]: I0929 19:33:23.210732 4741 scope.go:117] "RemoveContainer" containerID="915657e3007151e1d5101dbb2d89c3041033781338926c491ec154e4b25f1c9a" Sep 29 19:33:24 crc kubenswrapper[4741]: I0929 19:33:24.960552 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mtgjf" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="registry-server" containerID="cri-o://60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6" gracePeriod=2 Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.341482 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.505288 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j656r\" (UniqueName: \"kubernetes.io/projected/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-kube-api-access-j656r\") pod \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.505348 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-catalog-content\") pod \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.505489 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-utilities\") pod \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\" (UID: \"ccf6e879-2dee-4251-9752-8c7c1fcdeda0\") " Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.506676 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-utilities" (OuterVolumeSpecName: "utilities") pod "ccf6e879-2dee-4251-9752-8c7c1fcdeda0" (UID: "ccf6e879-2dee-4251-9752-8c7c1fcdeda0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.511275 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-kube-api-access-j656r" (OuterVolumeSpecName: "kube-api-access-j656r") pod "ccf6e879-2dee-4251-9752-8c7c1fcdeda0" (UID: "ccf6e879-2dee-4251-9752-8c7c1fcdeda0"). InnerVolumeSpecName "kube-api-access-j656r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.567613 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccf6e879-2dee-4251-9752-8c7c1fcdeda0" (UID: "ccf6e879-2dee-4251-9752-8c7c1fcdeda0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.607008 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.607049 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j656r\" (UniqueName: \"kubernetes.io/projected/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-kube-api-access-j656r\") on node \"crc\" DevicePath \"\"" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.607062 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf6e879-2dee-4251-9752-8c7c1fcdeda0-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.969825 4741 generic.go:334] "Generic (PLEG): container finished" podID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerID="60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6" exitCode=0 Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.969902 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtgjf" event={"ID":"ccf6e879-2dee-4251-9752-8c7c1fcdeda0","Type":"ContainerDied","Data":"60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6"} Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.969922 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtgjf" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.971027 4741 scope.go:117] "RemoveContainer" containerID="60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6" Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.971015 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtgjf" event={"ID":"ccf6e879-2dee-4251-9752-8c7c1fcdeda0","Type":"ContainerDied","Data":"0f0a8031e331c55605b990e1a44a93c2f1357ef877ea43866f1d47c47395ea37"} Sep 29 19:33:25 crc kubenswrapper[4741]: I0929 19:33:25.991233 4741 scope.go:117] "RemoveContainer" containerID="29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f" Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.009910 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtgjf"] Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.020104 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mtgjf"] Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.041179 4741 scope.go:117] "RemoveContainer" containerID="60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a" Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.056998 4741 scope.go:117] "RemoveContainer" containerID="60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6" Sep 29 19:33:26 crc kubenswrapper[4741]: E0929 19:33:26.057460 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6\": container with ID starting with 60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6 not found: ID does not exist" containerID="60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6" Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.057497 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6"} err="failed to get container status \"60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6\": rpc error: code = NotFound desc = could not find container \"60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6\": container with ID starting with 60769aef180f39959ef62c03cd06c8694cf65a1923a642f044a8381316aa4cf6 not found: ID does not exist" Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.057523 4741 scope.go:117] "RemoveContainer" containerID="29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f" Sep 29 19:33:26 crc kubenswrapper[4741]: E0929 19:33:26.057858 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f\": container with ID starting with 29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f not found: ID does not exist" containerID="29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f" Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.057901 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f"} err="failed to get container status \"29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f\": rpc error: code = NotFound desc = could not find container \"29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f\": container with ID starting with 29b2bc3f9275a0cd474f2cedc8cc8403858fb95a35fda34ed9205b05aaa9381f not found: ID does not exist" Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.057932 4741 scope.go:117] "RemoveContainer" containerID="60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a" Sep 29 19:33:26 crc kubenswrapper[4741]: E0929 19:33:26.058315 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a\": container with ID starting with 60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a not found: ID does not exist" containerID="60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a" Sep 29 19:33:26 crc kubenswrapper[4741]: I0929 19:33:26.058340 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a"} err="failed to get container status \"60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a\": rpc error: code = NotFound desc = could not find container \"60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a\": container with ID starting with 60c9e0e6dee1c1403c6052e3f05d9050a2bdbb34c9f8bedd8d54925a33d0ca9a not found: ID does not exist" Sep 29 19:33:27 crc kubenswrapper[4741]: I0929 19:33:27.095889 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" path="/var/lib/kubelet/pods/ccf6e879-2dee-4251-9752-8c7c1fcdeda0/volumes" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.487305 4741 scope.go:117] "RemoveContainer" containerID="fc163c8b456c4eae3276be79b996f9a990070e39bb81a710df286eea59af6f36" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.511838 4741 scope.go:117] "RemoveContainer" containerID="0e81cfe546352fd6f5ad379bfb7d977c1d18997906a11e352f7345acc20dfc48" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.543121 4741 scope.go:117] "RemoveContainer" containerID="443a8fa2069e99c030863c042183d939ecadb9561580d7eeef5f0284e38a1e60" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.593657 4741 scope.go:117] "RemoveContainer" containerID="5747543db2d4686ff0ce2d66484ba05733793d16bc7f1a3c6bf312799d425845" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.618042 4741 scope.go:117] "RemoveContainer" containerID="dfd2422ee682a56f33a264c928c2202071429b72837518805278e82828695484" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.638185 4741 scope.go:117] "RemoveContainer" containerID="bd272abdec76fda4950b811de1c1e048d2e279faf0d1fac0874af53e2d4b1e91" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.662923 4741 scope.go:117] "RemoveContainer" containerID="eac6f805571d33e47abcb8386b1e4a78317268ba124fe10a2e9376e94d5dd506" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.693947 4741 scope.go:117] "RemoveContainer" containerID="58644df2679ff5827ca80923ddee91fde9b4da2f9eb6e37ae982ca0c9572be55" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.714567 4741 scope.go:117] "RemoveContainer" containerID="c6c7dcc6a6a4840a78d37eb8074a4534402a4879ebd9e67c49333dcb254393dc" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.774078 4741 scope.go:117] "RemoveContainer" containerID="88822bc2858601e3da0fba4f6a6ca50dcd1d5c1fe288573a193d9458b638d155" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.800878 4741 scope.go:117] "RemoveContainer" containerID="bf70af070e1337a80f352e7891b4066ae9a350ffce11fe54650a0f5325a94c18" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.834022 4741 scope.go:117] "RemoveContainer" containerID="c7249b3b7837fc3520fcc2d531958323f2ab743d34b60e50c851a65328623db0" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.863294 4741 scope.go:117] "RemoveContainer" containerID="f0f5d8ecf900696782222753fa68359e26bb6b0234dc48e8dad493ca3f5967bc" Sep 29 19:34:23 crc kubenswrapper[4741]: I0929 19:34:23.900322 4741 scope.go:117] "RemoveContainer" containerID="4b30dc6f2664d0596e57ddb9855aca996f10f7caab7b0df97ee564ddcc5bc232" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.092859 4741 scope.go:117] "RemoveContainer" containerID="e1b73040b554cbd478a51c8b5816e58a0a68a53c90c12ff3ca2f8c2045c814bf" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.113176 4741 scope.go:117] "RemoveContainer" containerID="5adf7bb9a660c6c77e2e4c1e061e23a7547defbfb032ae8b851205e03cab997b" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.137752 4741 scope.go:117] "RemoveContainer" containerID="cce06dd4a20cab8469d6440e5d832029b66e459b52c5ff9b817bb86dbe5b5145" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.161022 4741 scope.go:117] "RemoveContainer" containerID="7e12e5f4e6b247101d296695c9a7418f70e077c57abdf6fe940893887187da15" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.199280 4741 scope.go:117] "RemoveContainer" containerID="58f9c958476dc4429b27a3cf8d2a755c71d7cd37a2d48d47dc954ccd9f7b2781" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.242427 4741 scope.go:117] "RemoveContainer" containerID="59e07c173f297fc52b58154fee10e2757e48b2c11cd92731bb4e2fc98f386bc7" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.260375 4741 scope.go:117] "RemoveContainer" containerID="9bac786058172af76d1feab8bbd593df29dab72cbf3ab45d0f20758962fff82c" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.279964 4741 scope.go:117] "RemoveContainer" containerID="73c176af015738a2be3c557c9aba6b6eb0d0e9b25133c488196d8f50acaa001c" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.298265 4741 scope.go:117] "RemoveContainer" containerID="5488046b5c71d7ada89c02641c087812ffa5561b258e4750ab91f6cd07383d28" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.322511 4741 scope.go:117] "RemoveContainer" containerID="56f048944583f6a7fea3651342397a37fb4283a7918ed222db829e37ecd52e27" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.372475 4741 scope.go:117] "RemoveContainer" containerID="b426eb53e2825c670095e43996bf58ecec4760a4bafd71a4d7304962625a63c7" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.390300 4741 scope.go:117] "RemoveContainer" containerID="c9909a6976e00256c9ff763bed1a932c9df0fd365ff4dabc64990d63e5be6cf4" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.406813 4741 scope.go:117] "RemoveContainer" containerID="3f85075833d16e2d3f3491a5e4ebd1422fbd5759d4cfda248972d88804bd4bf9" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.420038 4741 scope.go:117] "RemoveContainer" containerID="aaae2fd3fe24f56e30fb6d8a9846ff321b4277473b96cefb25838ece71c0bd8f" Sep 29 19:35:24 crc kubenswrapper[4741]: I0929 19:35:24.435360 4741 scope.go:117] "RemoveContainer" containerID="829444b757a521955e515b75492ea996bc9c06ad3fa6507cd00d66429d8cb77d" Sep 29 19:35:31 crc kubenswrapper[4741]: I0929 19:35:31.739296 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:35:31 crc kubenswrapper[4741]: I0929 19:35:31.739838 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:36:01 crc kubenswrapper[4741]: I0929 19:36:01.738756 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:36:01 crc kubenswrapper[4741]: I0929 19:36:01.739273 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:36:24 crc kubenswrapper[4741]: I0929 19:36:24.594181 4741 scope.go:117] "RemoveContainer" containerID="aa0bae00cb4dbdf6185261b12b6a457a4f38e5ab93463fa7e318709fcfe1a070" Sep 29 19:36:24 crc kubenswrapper[4741]: I0929 19:36:24.616065 4741 scope.go:117] "RemoveContainer" containerID="254ddac676d3d1212195f5f14061933f9ab70fa1f2be993950d36458f551f9d7" Sep 29 19:36:24 crc kubenswrapper[4741]: I0929 19:36:24.636763 4741 scope.go:117] "RemoveContainer" containerID="040aa5ec5b498aa5b60ba516e013b9839f5b611e8d7d5311e0be82ba0ff8b3c7" Sep 29 19:36:24 crc kubenswrapper[4741]: I0929 19:36:24.654891 4741 scope.go:117] "RemoveContainer" containerID="07ff09e353f0b9f0d4e0a433a86e107581fb646712047e332b67415755e5ec2d" Sep 29 19:36:24 crc kubenswrapper[4741]: I0929 19:36:24.689679 4741 scope.go:117] "RemoveContainer" containerID="1134aaa517963bd9f046f45533c4199e3a4bd021b68fd358ad1a47c8622b1634" Sep 29 19:36:24 crc kubenswrapper[4741]: I0929 19:36:24.706207 4741 scope.go:117] "RemoveContainer" containerID="8dccd7363f3ef4f5fc4ea08cd195ce440540c0c08f0954b35a076493bfa65bf4" Sep 29 19:36:24 crc kubenswrapper[4741]: I0929 19:36:24.723492 4741 scope.go:117] "RemoveContainer" containerID="e3ba9d353708ea132687e9ee72cfd9472209753189bc1863b2bf80c6d632fdd9" Sep 29 19:36:31 crc kubenswrapper[4741]: I0929 19:36:31.739314 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:36:31 crc kubenswrapper[4741]: I0929 19:36:31.739816 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:36:31 crc kubenswrapper[4741]: I0929 19:36:31.739849 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:36:31 crc kubenswrapper[4741]: I0929 19:36:31.740383 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:36:31 crc kubenswrapper[4741]: I0929 19:36:31.740458 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" gracePeriod=600 Sep 29 19:36:31 crc kubenswrapper[4741]: E0929 19:36:31.863207 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:36:32 crc kubenswrapper[4741]: I0929 19:36:32.575872 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" exitCode=0 Sep 29 19:36:32 crc kubenswrapper[4741]: I0929 19:36:32.575918 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8"} Sep 29 19:36:32 crc kubenswrapper[4741]: I0929 19:36:32.575961 4741 scope.go:117] "RemoveContainer" containerID="a962e46acb3fe13f6d108a85b06c3220bc7a353adc196e7c6597c33dec71e863" Sep 29 19:36:32 crc kubenswrapper[4741]: I0929 19:36:32.576446 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:36:32 crc kubenswrapper[4741]: E0929 19:36:32.576684 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:36:45 crc kubenswrapper[4741]: I0929 19:36:45.087746 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:36:45 crc kubenswrapper[4741]: E0929 19:36:45.088470 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:36:56 crc kubenswrapper[4741]: I0929 19:36:56.086358 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:36:56 crc kubenswrapper[4741]: E0929 19:36:56.087027 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:37:07 crc kubenswrapper[4741]: I0929 19:37:07.085569 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:37:07 crc kubenswrapper[4741]: E0929 19:37:07.086799 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:37:18 crc kubenswrapper[4741]: I0929 19:37:18.086232 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:37:18 crc kubenswrapper[4741]: E0929 19:37:18.087216 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:37:24 crc kubenswrapper[4741]: I0929 19:37:24.825071 4741 scope.go:117] "RemoveContainer" containerID="aca50d7d0a09ea3de4b166aafd5c60114abc7d8e450c9b107c4c352810e4b20a" Sep 29 19:37:24 crc kubenswrapper[4741]: I0929 19:37:24.849217 4741 scope.go:117] "RemoveContainer" containerID="d48e3684ebfff17addc725043e0493445f56a413b0de369662ab33c743879d3d" Sep 29 19:37:31 crc kubenswrapper[4741]: I0929 19:37:31.086414 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:37:31 crc kubenswrapper[4741]: E0929 19:37:31.088174 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:37:42 crc kubenswrapper[4741]: I0929 19:37:42.086083 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:37:42 crc kubenswrapper[4741]: E0929 19:37:42.087746 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:37:53 crc kubenswrapper[4741]: I0929 19:37:53.085706 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:37:53 crc kubenswrapper[4741]: E0929 19:37:53.088349 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:38:07 crc kubenswrapper[4741]: I0929 19:38:07.086508 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:38:07 crc kubenswrapper[4741]: E0929 19:38:07.087159 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:38:20 crc kubenswrapper[4741]: I0929 19:38:20.086069 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:38:20 crc kubenswrapper[4741]: E0929 19:38:20.086696 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:38:34 crc kubenswrapper[4741]: I0929 19:38:34.085957 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:38:34 crc kubenswrapper[4741]: E0929 19:38:34.086743 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:38:46 crc kubenswrapper[4741]: I0929 19:38:46.086277 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:38:46 crc kubenswrapper[4741]: E0929 19:38:46.087607 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:38:59 crc kubenswrapper[4741]: I0929 19:38:59.089682 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:38:59 crc kubenswrapper[4741]: E0929 19:38:59.090379 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:39:14 crc kubenswrapper[4741]: I0929 19:39:14.088321 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:39:14 crc kubenswrapper[4741]: E0929 19:39:14.090144 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:39:29 crc kubenswrapper[4741]: I0929 19:39:29.094072 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:39:29 crc kubenswrapper[4741]: E0929 19:39:29.094749 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:39:41 crc kubenswrapper[4741]: I0929 19:39:41.086983 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:39:41 crc kubenswrapper[4741]: E0929 19:39:41.088289 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.199933 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jwgh7"] Sep 29 19:39:42 crc kubenswrapper[4741]: E0929 19:39:42.201525 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="extract-utilities" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.201556 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="extract-utilities" Sep 29 19:39:42 crc kubenswrapper[4741]: E0929 19:39:42.201592 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="registry-server" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.201605 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="registry-server" Sep 29 19:39:42 crc kubenswrapper[4741]: E0929 19:39:42.201644 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="extract-content" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.201657 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="extract-content" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.202625 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf6e879-2dee-4251-9752-8c7c1fcdeda0" containerName="registry-server" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.211376 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.224666 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jwgh7"] Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.335085 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkvtq\" (UniqueName: \"kubernetes.io/projected/5e4824c0-17f4-40ea-814c-2bc6783c3df9-kube-api-access-gkvtq\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.335455 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-utilities\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.335546 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-catalog-content\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.436379 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkvtq\" (UniqueName: \"kubernetes.io/projected/5e4824c0-17f4-40ea-814c-2bc6783c3df9-kube-api-access-gkvtq\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.436472 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-utilities\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.436496 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-catalog-content\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.437015 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-catalog-content\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.437614 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-utilities\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.468646 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkvtq\" (UniqueName: \"kubernetes.io/projected/5e4824c0-17f4-40ea-814c-2bc6783c3df9-kube-api-access-gkvtq\") pod \"redhat-operators-jwgh7\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.531971 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:42 crc kubenswrapper[4741]: I0929 19:39:42.978590 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jwgh7"] Sep 29 19:39:43 crc kubenswrapper[4741]: I0929 19:39:43.140608 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerStarted","Data":"d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e"} Sep 29 19:39:43 crc kubenswrapper[4741]: I0929 19:39:43.141266 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerStarted","Data":"511bb9294132b837d671021548ad77bb366cc2bf1db7abf063161d068a7e9d4b"} Sep 29 19:39:43 crc kubenswrapper[4741]: I0929 19:39:43.142583 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 19:39:44 crc kubenswrapper[4741]: I0929 19:39:44.148671 4741 generic.go:334] "Generic (PLEG): container finished" podID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerID="d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e" exitCode=0 Sep 29 19:39:44 crc kubenswrapper[4741]: I0929 19:39:44.148728 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerDied","Data":"d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e"} Sep 29 19:39:44 crc kubenswrapper[4741]: I0929 19:39:44.148935 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerStarted","Data":"2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2"} Sep 29 19:39:45 crc kubenswrapper[4741]: I0929 19:39:45.160070 4741 generic.go:334] "Generic (PLEG): container finished" podID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerID="2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2" exitCode=0 Sep 29 19:39:45 crc kubenswrapper[4741]: I0929 19:39:45.160123 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerDied","Data":"2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2"} Sep 29 19:39:46 crc kubenswrapper[4741]: I0929 19:39:46.170077 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerStarted","Data":"aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530"} Sep 29 19:39:46 crc kubenswrapper[4741]: I0929 19:39:46.190139 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jwgh7" podStartSLOduration=1.547586803 podStartE2EDuration="4.190121971s" podCreationTimestamp="2025-09-29 19:39:42 +0000 UTC" firstStartedPulling="2025-09-29 19:39:43.142312202 +0000 UTC m=+1824.790101534" lastFinishedPulling="2025-09-29 19:39:45.78484736 +0000 UTC m=+1827.432636702" observedRunningTime="2025-09-29 19:39:46.186280311 +0000 UTC m=+1827.834069643" watchObservedRunningTime="2025-09-29 19:39:46.190121971 +0000 UTC m=+1827.837911303" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.386985 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t7tnr"] Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.389484 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.393776 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7tnr"] Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.500761 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-utilities\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.500878 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmj7m\" (UniqueName: \"kubernetes.io/projected/a008845b-fc65-4c1b-82e6-cb860e622034-kube-api-access-jmj7m\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.501013 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-catalog-content\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.601970 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmj7m\" (UniqueName: \"kubernetes.io/projected/a008845b-fc65-4c1b-82e6-cb860e622034-kube-api-access-jmj7m\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.602075 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-catalog-content\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.602131 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-utilities\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.602763 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-utilities\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.603147 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-catalog-content\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.623601 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmj7m\" (UniqueName: \"kubernetes.io/projected/a008845b-fc65-4c1b-82e6-cb860e622034-kube-api-access-jmj7m\") pod \"certified-operators-t7tnr\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:47 crc kubenswrapper[4741]: I0929 19:39:47.746986 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:48 crc kubenswrapper[4741]: I0929 19:39:48.011074 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7tnr"] Sep 29 19:39:48 crc kubenswrapper[4741]: W0929 19:39:48.026078 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda008845b_fc65_4c1b_82e6_cb860e622034.slice/crio-76cf771169f6c50bd1154790ae15f5b9e00d6fb983f1527260a2955c684cd61c WatchSource:0}: Error finding container 76cf771169f6c50bd1154790ae15f5b9e00d6fb983f1527260a2955c684cd61c: Status 404 returned error can't find the container with id 76cf771169f6c50bd1154790ae15f5b9e00d6fb983f1527260a2955c684cd61c Sep 29 19:39:48 crc kubenswrapper[4741]: I0929 19:39:48.197847 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerStarted","Data":"ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c"} Sep 29 19:39:48 crc kubenswrapper[4741]: I0929 19:39:48.198246 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerStarted","Data":"76cf771169f6c50bd1154790ae15f5b9e00d6fb983f1527260a2955c684cd61c"} Sep 29 19:39:49 crc kubenswrapper[4741]: I0929 19:39:49.205838 4741 generic.go:334] "Generic (PLEG): container finished" podID="a008845b-fc65-4c1b-82e6-cb860e622034" containerID="ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c" exitCode=0 Sep 29 19:39:49 crc kubenswrapper[4741]: I0929 19:39:49.205891 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerDied","Data":"ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c"} Sep 29 19:39:50 crc kubenswrapper[4741]: I0929 19:39:50.213780 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerStarted","Data":"98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e"} Sep 29 19:39:51 crc kubenswrapper[4741]: I0929 19:39:51.227363 4741 generic.go:334] "Generic (PLEG): container finished" podID="a008845b-fc65-4c1b-82e6-cb860e622034" containerID="98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e" exitCode=0 Sep 29 19:39:51 crc kubenswrapper[4741]: I0929 19:39:51.227568 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerDied","Data":"98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e"} Sep 29 19:39:52 crc kubenswrapper[4741]: I0929 19:39:52.236803 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerStarted","Data":"0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499"} Sep 29 19:39:52 crc kubenswrapper[4741]: I0929 19:39:52.273106 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t7tnr" podStartSLOduration=2.741817472 podStartE2EDuration="5.273086922s" podCreationTimestamp="2025-09-29 19:39:47 +0000 UTC" firstStartedPulling="2025-09-29 19:39:49.207590654 +0000 UTC m=+1830.855380006" lastFinishedPulling="2025-09-29 19:39:51.738860114 +0000 UTC m=+1833.386649456" observedRunningTime="2025-09-29 19:39:52.268103986 +0000 UTC m=+1833.915893328" watchObservedRunningTime="2025-09-29 19:39:52.273086922 +0000 UTC m=+1833.920876254" Sep 29 19:39:52 crc kubenswrapper[4741]: I0929 19:39:52.533338 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:52 crc kubenswrapper[4741]: I0929 19:39:52.533431 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:52 crc kubenswrapper[4741]: I0929 19:39:52.572286 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:53 crc kubenswrapper[4741]: I0929 19:39:53.319796 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:55 crc kubenswrapper[4741]: I0929 19:39:55.085915 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:39:55 crc kubenswrapper[4741]: E0929 19:39:55.086230 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.183001 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jwgh7"] Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.183907 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jwgh7" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="registry-server" containerID="cri-o://aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530" gracePeriod=2 Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.617276 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.732013 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-catalog-content\") pod \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.732125 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkvtq\" (UniqueName: \"kubernetes.io/projected/5e4824c0-17f4-40ea-814c-2bc6783c3df9-kube-api-access-gkvtq\") pod \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.732214 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-utilities\") pod \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\" (UID: \"5e4824c0-17f4-40ea-814c-2bc6783c3df9\") " Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.733373 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-utilities" (OuterVolumeSpecName: "utilities") pod "5e4824c0-17f4-40ea-814c-2bc6783c3df9" (UID: "5e4824c0-17f4-40ea-814c-2bc6783c3df9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.737786 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4824c0-17f4-40ea-814c-2bc6783c3df9-kube-api-access-gkvtq" (OuterVolumeSpecName: "kube-api-access-gkvtq") pod "5e4824c0-17f4-40ea-814c-2bc6783c3df9" (UID: "5e4824c0-17f4-40ea-814c-2bc6783c3df9"). InnerVolumeSpecName "kube-api-access-gkvtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.834114 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.834163 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkvtq\" (UniqueName: \"kubernetes.io/projected/5e4824c0-17f4-40ea-814c-2bc6783c3df9-kube-api-access-gkvtq\") on node \"crc\" DevicePath \"\"" Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.843650 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e4824c0-17f4-40ea-814c-2bc6783c3df9" (UID: "5e4824c0-17f4-40ea-814c-2bc6783c3df9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:39:56 crc kubenswrapper[4741]: I0929 19:39:56.936474 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4824c0-17f4-40ea-814c-2bc6783c3df9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.288997 4741 generic.go:334] "Generic (PLEG): container finished" podID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerID="aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530" exitCode=0 Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.289061 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerDied","Data":"aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530"} Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.289101 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwgh7" event={"ID":"5e4824c0-17f4-40ea-814c-2bc6783c3df9","Type":"ContainerDied","Data":"511bb9294132b837d671021548ad77bb366cc2bf1db7abf063161d068a7e9d4b"} Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.289129 4741 scope.go:117] "RemoveContainer" containerID="aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.289301 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwgh7" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.311946 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jwgh7"] Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.317062 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jwgh7"] Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.327075 4741 scope.go:117] "RemoveContainer" containerID="2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.351866 4741 scope.go:117] "RemoveContainer" containerID="d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.382024 4741 scope.go:117] "RemoveContainer" containerID="aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530" Sep 29 19:39:57 crc kubenswrapper[4741]: E0929 19:39:57.383028 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530\": container with ID starting with aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530 not found: ID does not exist" containerID="aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.383086 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530"} err="failed to get container status \"aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530\": rpc error: code = NotFound desc = could not find container \"aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530\": container with ID starting with aa2c71c62fa1cb72407f4efbc2a9d211ba5abe1eeb3531cf1ca1b05de27ee530 not found: ID does not exist" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.383122 4741 scope.go:117] "RemoveContainer" containerID="2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2" Sep 29 19:39:57 crc kubenswrapper[4741]: E0929 19:39:57.383483 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2\": container with ID starting with 2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2 not found: ID does not exist" containerID="2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.383518 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2"} err="failed to get container status \"2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2\": rpc error: code = NotFound desc = could not find container \"2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2\": container with ID starting with 2c588b7fd8b0a7f8020c68d8b447f5769885148a98941e0720c2071e39715cd2 not found: ID does not exist" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.383544 4741 scope.go:117] "RemoveContainer" containerID="d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e" Sep 29 19:39:57 crc kubenswrapper[4741]: E0929 19:39:57.383928 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e\": container with ID starting with d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e not found: ID does not exist" containerID="d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.383959 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e"} err="failed to get container status \"d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e\": rpc error: code = NotFound desc = could not find container \"d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e\": container with ID starting with d9fff8198d7dfe66a3800bd9977bd58fc79ac8860a928a67f1bdd3e5a40dae0e not found: ID does not exist" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.747576 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.747885 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:57 crc kubenswrapper[4741]: I0929 19:39:57.799100 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:58 crc kubenswrapper[4741]: I0929 19:39:58.339183 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:39:59 crc kubenswrapper[4741]: I0929 19:39:59.096857 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" path="/var/lib/kubelet/pods/5e4824c0-17f4-40ea-814c-2bc6783c3df9/volumes" Sep 29 19:39:59 crc kubenswrapper[4741]: I0929 19:39:59.983696 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7tnr"] Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.326843 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t7tnr" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="registry-server" containerID="cri-o://0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499" gracePeriod=2 Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.744136 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.907515 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmj7m\" (UniqueName: \"kubernetes.io/projected/a008845b-fc65-4c1b-82e6-cb860e622034-kube-api-access-jmj7m\") pod \"a008845b-fc65-4c1b-82e6-cb860e622034\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.907622 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-catalog-content\") pod \"a008845b-fc65-4c1b-82e6-cb860e622034\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.907703 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-utilities\") pod \"a008845b-fc65-4c1b-82e6-cb860e622034\" (UID: \"a008845b-fc65-4c1b-82e6-cb860e622034\") " Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.908648 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-utilities" (OuterVolumeSpecName: "utilities") pod "a008845b-fc65-4c1b-82e6-cb860e622034" (UID: "a008845b-fc65-4c1b-82e6-cb860e622034"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.915713 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a008845b-fc65-4c1b-82e6-cb860e622034-kube-api-access-jmj7m" (OuterVolumeSpecName: "kube-api-access-jmj7m") pod "a008845b-fc65-4c1b-82e6-cb860e622034" (UID: "a008845b-fc65-4c1b-82e6-cb860e622034"). InnerVolumeSpecName "kube-api-access-jmj7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:40:01 crc kubenswrapper[4741]: I0929 19:40:01.961894 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a008845b-fc65-4c1b-82e6-cb860e622034" (UID: "a008845b-fc65-4c1b-82e6-cb860e622034"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.009453 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.009486 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008845b-fc65-4c1b-82e6-cb860e622034-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.009498 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmj7m\" (UniqueName: \"kubernetes.io/projected/a008845b-fc65-4c1b-82e6-cb860e622034-kube-api-access-jmj7m\") on node \"crc\" DevicePath \"\"" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.337724 4741 generic.go:334] "Generic (PLEG): container finished" podID="a008845b-fc65-4c1b-82e6-cb860e622034" containerID="0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499" exitCode=0 Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.337769 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7tnr" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.337788 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerDied","Data":"0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499"} Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.338159 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7tnr" event={"ID":"a008845b-fc65-4c1b-82e6-cb860e622034","Type":"ContainerDied","Data":"76cf771169f6c50bd1154790ae15f5b9e00d6fb983f1527260a2955c684cd61c"} Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.338189 4741 scope.go:117] "RemoveContainer" containerID="0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.358972 4741 scope.go:117] "RemoveContainer" containerID="98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.372852 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7tnr"] Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.378148 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t7tnr"] Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.407045 4741 scope.go:117] "RemoveContainer" containerID="ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.424523 4741 scope.go:117] "RemoveContainer" containerID="0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499" Sep 29 19:40:02 crc kubenswrapper[4741]: E0929 19:40:02.424817 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499\": container with ID starting with 0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499 not found: ID does not exist" containerID="0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.424862 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499"} err="failed to get container status \"0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499\": rpc error: code = NotFound desc = could not find container \"0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499\": container with ID starting with 0ea0d4289e014f6fccb436d9bfac4bf6ea30fb5d598ca4c3b6c0451c731f4499 not found: ID does not exist" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.424886 4741 scope.go:117] "RemoveContainer" containerID="98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e" Sep 29 19:40:02 crc kubenswrapper[4741]: E0929 19:40:02.425228 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e\": container with ID starting with 98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e not found: ID does not exist" containerID="98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.425257 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e"} err="failed to get container status \"98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e\": rpc error: code = NotFound desc = could not find container \"98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e\": container with ID starting with 98f364aa8d559041cbb0f660cd12c1d901dbdd60bfb6120a93d6cb258f2f9a4e not found: ID does not exist" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.425274 4741 scope.go:117] "RemoveContainer" containerID="ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c" Sep 29 19:40:02 crc kubenswrapper[4741]: E0929 19:40:02.425660 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c\": container with ID starting with ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c not found: ID does not exist" containerID="ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c" Sep 29 19:40:02 crc kubenswrapper[4741]: I0929 19:40:02.425684 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c"} err="failed to get container status \"ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c\": rpc error: code = NotFound desc = could not find container \"ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c\": container with ID starting with ebaefcb2599e7a5eb6fc42fb09d41e508ba00f637f2141d7395b3c809707591c not found: ID does not exist" Sep 29 19:40:03 crc kubenswrapper[4741]: I0929 19:40:03.095264 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" path="/var/lib/kubelet/pods/a008845b-fc65-4c1b-82e6-cb860e622034/volumes" Sep 29 19:40:09 crc kubenswrapper[4741]: I0929 19:40:09.091714 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:40:09 crc kubenswrapper[4741]: E0929 19:40:09.092494 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:40:24 crc kubenswrapper[4741]: I0929 19:40:24.086411 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:40:24 crc kubenswrapper[4741]: E0929 19:40:24.087201 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:40:37 crc kubenswrapper[4741]: I0929 19:40:37.086351 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:40:37 crc kubenswrapper[4741]: E0929 19:40:37.087058 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:40:48 crc kubenswrapper[4741]: I0929 19:40:48.085733 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:40:48 crc kubenswrapper[4741]: E0929 19:40:48.086478 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:41:01 crc kubenswrapper[4741]: I0929 19:41:01.086774 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:41:01 crc kubenswrapper[4741]: E0929 19:41:01.087688 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:41:15 crc kubenswrapper[4741]: I0929 19:41:15.086101 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:41:15 crc kubenswrapper[4741]: E0929 19:41:15.087078 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:41:28 crc kubenswrapper[4741]: I0929 19:41:28.086243 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:41:28 crc kubenswrapper[4741]: E0929 19:41:28.087092 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:41:40 crc kubenswrapper[4741]: I0929 19:41:40.086061 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:41:41 crc kubenswrapper[4741]: I0929 19:41:41.145335 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"3a8181336d2441af20e9c2452f4c8711ec838be48460263fc0c16a5200c4fdcf"} Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.462672 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kfvnc"] Sep 29 19:42:53 crc kubenswrapper[4741]: E0929 19:42:53.463843 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="extract-content" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.463860 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="extract-content" Sep 29 19:42:53 crc kubenswrapper[4741]: E0929 19:42:53.463878 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="extract-content" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.463885 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="extract-content" Sep 29 19:42:53 crc kubenswrapper[4741]: E0929 19:42:53.463910 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="extract-utilities" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.463919 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="extract-utilities" Sep 29 19:42:53 crc kubenswrapper[4741]: E0929 19:42:53.463934 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="registry-server" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.463941 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="registry-server" Sep 29 19:42:53 crc kubenswrapper[4741]: E0929 19:42:53.463963 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="extract-utilities" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.463970 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="extract-utilities" Sep 29 19:42:53 crc kubenswrapper[4741]: E0929 19:42:53.463990 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="registry-server" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.464000 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="registry-server" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.464201 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a008845b-fc65-4c1b-82e6-cb860e622034" containerName="registry-server" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.464227 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4824c0-17f4-40ea-814c-2bc6783c3df9" containerName="registry-server" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.465839 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.468108 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfvnc"] Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.632218 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67hwm\" (UniqueName: \"kubernetes.io/projected/a95ff214-ff50-41a6-b339-0d6fcd17ab82-kube-api-access-67hwm\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.632529 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-utilities\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.632593 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-catalog-content\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.733656 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-utilities\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.733704 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-catalog-content\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.733768 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67hwm\" (UniqueName: \"kubernetes.io/projected/a95ff214-ff50-41a6-b339-0d6fcd17ab82-kube-api-access-67hwm\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.734258 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-utilities\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.734296 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-catalog-content\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.751349 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67hwm\" (UniqueName: \"kubernetes.io/projected/a95ff214-ff50-41a6-b339-0d6fcd17ab82-kube-api-access-67hwm\") pod \"redhat-marketplace-kfvnc\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:53 crc kubenswrapper[4741]: I0929 19:42:53.787956 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:42:54 crc kubenswrapper[4741]: I0929 19:42:54.185589 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfvnc"] Sep 29 19:42:54 crc kubenswrapper[4741]: I0929 19:42:54.712065 4741 generic.go:334] "Generic (PLEG): container finished" podID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerID="857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5" exitCode=0 Sep 29 19:42:54 crc kubenswrapper[4741]: I0929 19:42:54.712151 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfvnc" event={"ID":"a95ff214-ff50-41a6-b339-0d6fcd17ab82","Type":"ContainerDied","Data":"857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5"} Sep 29 19:42:54 crc kubenswrapper[4741]: I0929 19:42:54.712655 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfvnc" event={"ID":"a95ff214-ff50-41a6-b339-0d6fcd17ab82","Type":"ContainerStarted","Data":"e1dd06732a2f75972d07389d449b2a47f399c77a6edd2cbbfe14055128c509c3"} Sep 29 19:42:55 crc kubenswrapper[4741]: I0929 19:42:55.721433 4741 generic.go:334] "Generic (PLEG): container finished" podID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerID="214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385" exitCode=0 Sep 29 19:42:55 crc kubenswrapper[4741]: I0929 19:42:55.721614 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfvnc" event={"ID":"a95ff214-ff50-41a6-b339-0d6fcd17ab82","Type":"ContainerDied","Data":"214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385"} Sep 29 19:42:56 crc kubenswrapper[4741]: I0929 19:42:56.730874 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfvnc" event={"ID":"a95ff214-ff50-41a6-b339-0d6fcd17ab82","Type":"ContainerStarted","Data":"19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83"} Sep 29 19:42:56 crc kubenswrapper[4741]: I0929 19:42:56.750033 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kfvnc" podStartSLOduration=2.27306981 podStartE2EDuration="3.750011719s" podCreationTimestamp="2025-09-29 19:42:53 +0000 UTC" firstStartedPulling="2025-09-29 19:42:54.713799072 +0000 UTC m=+2016.361588394" lastFinishedPulling="2025-09-29 19:42:56.190740961 +0000 UTC m=+2017.838530303" observedRunningTime="2025-09-29 19:42:56.745842098 +0000 UTC m=+2018.393631420" watchObservedRunningTime="2025-09-29 19:42:56.750011719 +0000 UTC m=+2018.397801051" Sep 29 19:43:03 crc kubenswrapper[4741]: I0929 19:43:03.788974 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:43:03 crc kubenswrapper[4741]: I0929 19:43:03.789537 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:43:03 crc kubenswrapper[4741]: I0929 19:43:03.830153 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:43:04 crc kubenswrapper[4741]: I0929 19:43:04.836991 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:43:04 crc kubenswrapper[4741]: I0929 19:43:04.879462 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfvnc"] Sep 29 19:43:06 crc kubenswrapper[4741]: I0929 19:43:06.805550 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kfvnc" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="registry-server" containerID="cri-o://19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83" gracePeriod=2 Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.162556 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.323544 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-catalog-content\") pod \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.323619 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-utilities\") pod \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.324633 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67hwm\" (UniqueName: \"kubernetes.io/projected/a95ff214-ff50-41a6-b339-0d6fcd17ab82-kube-api-access-67hwm\") pod \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\" (UID: \"a95ff214-ff50-41a6-b339-0d6fcd17ab82\") " Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.324907 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-utilities" (OuterVolumeSpecName: "utilities") pod "a95ff214-ff50-41a6-b339-0d6fcd17ab82" (UID: "a95ff214-ff50-41a6-b339-0d6fcd17ab82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.325834 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.329151 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a95ff214-ff50-41a6-b339-0d6fcd17ab82-kube-api-access-67hwm" (OuterVolumeSpecName: "kube-api-access-67hwm") pod "a95ff214-ff50-41a6-b339-0d6fcd17ab82" (UID: "a95ff214-ff50-41a6-b339-0d6fcd17ab82"). InnerVolumeSpecName "kube-api-access-67hwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.337677 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a95ff214-ff50-41a6-b339-0d6fcd17ab82" (UID: "a95ff214-ff50-41a6-b339-0d6fcd17ab82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.427654 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67hwm\" (UniqueName: \"kubernetes.io/projected/a95ff214-ff50-41a6-b339-0d6fcd17ab82-kube-api-access-67hwm\") on node \"crc\" DevicePath \"\"" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.427694 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a95ff214-ff50-41a6-b339-0d6fcd17ab82-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.812313 4741 generic.go:334] "Generic (PLEG): container finished" podID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerID="19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83" exitCode=0 Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.812361 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kfvnc" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.812358 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfvnc" event={"ID":"a95ff214-ff50-41a6-b339-0d6fcd17ab82","Type":"ContainerDied","Data":"19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83"} Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.812522 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kfvnc" event={"ID":"a95ff214-ff50-41a6-b339-0d6fcd17ab82","Type":"ContainerDied","Data":"e1dd06732a2f75972d07389d449b2a47f399c77a6edd2cbbfe14055128c509c3"} Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.812558 4741 scope.go:117] "RemoveContainer" containerID="19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.832552 4741 scope.go:117] "RemoveContainer" containerID="214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.842252 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfvnc"] Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.847686 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kfvnc"] Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.864295 4741 scope.go:117] "RemoveContainer" containerID="857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.878452 4741 scope.go:117] "RemoveContainer" containerID="19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83" Sep 29 19:43:07 crc kubenswrapper[4741]: E0929 19:43:07.878927 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83\": container with ID starting with 19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83 not found: ID does not exist" containerID="19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.878958 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83"} err="failed to get container status \"19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83\": rpc error: code = NotFound desc = could not find container \"19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83\": container with ID starting with 19562a6021511ab6e69a6fe926bd2333c97e1557b18966af0a367d88a3ae7b83 not found: ID does not exist" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.878981 4741 scope.go:117] "RemoveContainer" containerID="214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385" Sep 29 19:43:07 crc kubenswrapper[4741]: E0929 19:43:07.879268 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385\": container with ID starting with 214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385 not found: ID does not exist" containerID="214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.879444 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385"} err="failed to get container status \"214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385\": rpc error: code = NotFound desc = could not find container \"214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385\": container with ID starting with 214ed36fbd4eff3d443a7ff8893b8345e9d2057324e9141939095360d0b0a385 not found: ID does not exist" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.879553 4741 scope.go:117] "RemoveContainer" containerID="857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5" Sep 29 19:43:07 crc kubenswrapper[4741]: E0929 19:43:07.879903 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5\": container with ID starting with 857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5 not found: ID does not exist" containerID="857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5" Sep 29 19:43:07 crc kubenswrapper[4741]: I0929 19:43:07.879931 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5"} err="failed to get container status \"857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5\": rpc error: code = NotFound desc = could not find container \"857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5\": container with ID starting with 857e81229a41964fdad78d0cf582377c6c89d98952db85be85182226ecf2fde5 not found: ID does not exist" Sep 29 19:43:09 crc kubenswrapper[4741]: I0929 19:43:09.094039 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" path="/var/lib/kubelet/pods/a95ff214-ff50-41a6-b339-0d6fcd17ab82/volumes" Sep 29 19:44:01 crc kubenswrapper[4741]: I0929 19:44:01.738576 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:44:01 crc kubenswrapper[4741]: I0929 19:44:01.739191 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:44:31 crc kubenswrapper[4741]: I0929 19:44:31.739494 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:44:31 crc kubenswrapper[4741]: I0929 19:44:31.740139 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.163748 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc"] Sep 29 19:45:00 crc kubenswrapper[4741]: E0929 19:45:00.164829 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="extract-utilities" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.164851 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="extract-utilities" Sep 29 19:45:00 crc kubenswrapper[4741]: E0929 19:45:00.165464 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="extract-content" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.165486 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="extract-content" Sep 29 19:45:00 crc kubenswrapper[4741]: E0929 19:45:00.165515 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="registry-server" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.165528 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="registry-server" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.165823 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a95ff214-ff50-41a6-b339-0d6fcd17ab82" containerName="registry-server" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.166839 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.169116 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.170861 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.174384 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc"] Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.276894 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-config-volume\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.277027 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-secret-volume\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.277092 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrmtz\" (UniqueName: \"kubernetes.io/projected/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-kube-api-access-hrmtz\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.378249 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-secret-volume\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.378323 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrmtz\" (UniqueName: \"kubernetes.io/projected/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-kube-api-access-hrmtz\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.378358 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-config-volume\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.379295 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-config-volume\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.389095 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-secret-volume\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.394261 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrmtz\" (UniqueName: \"kubernetes.io/projected/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-kube-api-access-hrmtz\") pod \"collect-profiles-29319585-v42tc\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.492083 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:00 crc kubenswrapper[4741]: W0929 19:45:00.925696 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfe76ac4_2e3d_4366_9ce6_075ff7eb946e.slice/crio-b46000dd4392161839887552c276dc1251d914f6b6aecee68dfa147131f06311 WatchSource:0}: Error finding container b46000dd4392161839887552c276dc1251d914f6b6aecee68dfa147131f06311: Status 404 returned error can't find the container with id b46000dd4392161839887552c276dc1251d914f6b6aecee68dfa147131f06311 Sep 29 19:45:00 crc kubenswrapper[4741]: I0929 19:45:00.927129 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc"] Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.678211 4741 generic.go:334] "Generic (PLEG): container finished" podID="dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" containerID="fd0c67f59704a3a436a57d65109ce936b05199316e906b69d938fd74e4e423b0" exitCode=0 Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.678503 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" event={"ID":"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e","Type":"ContainerDied","Data":"fd0c67f59704a3a436a57d65109ce936b05199316e906b69d938fd74e4e423b0"} Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.678537 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" event={"ID":"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e","Type":"ContainerStarted","Data":"b46000dd4392161839887552c276dc1251d914f6b6aecee68dfa147131f06311"} Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.738479 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.738533 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.738567 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.739074 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a8181336d2441af20e9c2452f4c8711ec838be48460263fc0c16a5200c4fdcf"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:45:01 crc kubenswrapper[4741]: I0929 19:45:01.739122 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://3a8181336d2441af20e9c2452f4c8711ec838be48460263fc0c16a5200c4fdcf" gracePeriod=600 Sep 29 19:45:02 crc kubenswrapper[4741]: I0929 19:45:02.694711 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="3a8181336d2441af20e9c2452f4c8711ec838be48460263fc0c16a5200c4fdcf" exitCode=0 Sep 29 19:45:02 crc kubenswrapper[4741]: I0929 19:45:02.694787 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"3a8181336d2441af20e9c2452f4c8711ec838be48460263fc0c16a5200c4fdcf"} Sep 29 19:45:02 crc kubenswrapper[4741]: I0929 19:45:02.695101 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff"} Sep 29 19:45:02 crc kubenswrapper[4741]: I0929 19:45:02.695123 4741 scope.go:117] "RemoveContainer" containerID="c32a739de70886f7839fa09103a2ee1ee98efc75a1d1995c0cd8c84cb6b822f8" Sep 29 19:45:02 crc kubenswrapper[4741]: I0929 19:45:02.985176 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.120643 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrmtz\" (UniqueName: \"kubernetes.io/projected/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-kube-api-access-hrmtz\") pod \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.120751 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-config-volume\") pod \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.120839 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-secret-volume\") pod \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\" (UID: \"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e\") " Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.121585 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-config-volume" (OuterVolumeSpecName: "config-volume") pod "dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" (UID: "dfe76ac4-2e3d-4366-9ce6-075ff7eb946e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.126819 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-kube-api-access-hrmtz" (OuterVolumeSpecName: "kube-api-access-hrmtz") pod "dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" (UID: "dfe76ac4-2e3d-4366-9ce6-075ff7eb946e"). InnerVolumeSpecName "kube-api-access-hrmtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.130574 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" (UID: "dfe76ac4-2e3d-4366-9ce6-075ff7eb946e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.223835 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.223914 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrmtz\" (UniqueName: \"kubernetes.io/projected/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-kube-api-access-hrmtz\") on node \"crc\" DevicePath \"\"" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.223923 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.710797 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" event={"ID":"dfe76ac4-2e3d-4366-9ce6-075ff7eb946e","Type":"ContainerDied","Data":"b46000dd4392161839887552c276dc1251d914f6b6aecee68dfa147131f06311"} Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.711166 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b46000dd4392161839887552c276dc1251d914f6b6aecee68dfa147131f06311" Sep 29 19:45:03 crc kubenswrapper[4741]: I0929 19:45:03.711236 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc" Sep 29 19:45:04 crc kubenswrapper[4741]: I0929 19:45:04.067793 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk"] Sep 29 19:45:04 crc kubenswrapper[4741]: I0929 19:45:04.072452 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319540-qf6zk"] Sep 29 19:45:05 crc kubenswrapper[4741]: I0929 19:45:05.100783 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6394941-7bc5-45db-9404-6fa52d0fb10f" path="/var/lib/kubelet/pods/b6394941-7bc5-45db-9404-6fa52d0fb10f/volumes" Sep 29 19:45:25 crc kubenswrapper[4741]: I0929 19:45:25.070505 4741 scope.go:117] "RemoveContainer" containerID="292180dc73a15c3eab4c0cf1929f8f62499d6f33cd171557c9522193af31e653" Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.822571 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q4dbl"] Sep 29 19:45:30 crc kubenswrapper[4741]: E0929 19:45:30.823707 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" containerName="collect-profiles" Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.823726 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" containerName="collect-profiles" Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.823975 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" containerName="collect-profiles" Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.825486 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.828486 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q4dbl"] Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.957311 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-utilities\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.957652 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qnxk\" (UniqueName: \"kubernetes.io/projected/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-kube-api-access-6qnxk\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:30 crc kubenswrapper[4741]: I0929 19:45:30.957671 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-catalog-content\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.059145 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-utilities\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.059231 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qnxk\" (UniqueName: \"kubernetes.io/projected/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-kube-api-access-6qnxk\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.059252 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-catalog-content\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.059740 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-catalog-content\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.059874 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-utilities\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.087509 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qnxk\" (UniqueName: \"kubernetes.io/projected/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-kube-api-access-6qnxk\") pod \"community-operators-q4dbl\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.157052 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.622729 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q4dbl"] Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.934027 4741 generic.go:334] "Generic (PLEG): container finished" podID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerID="20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29" exitCode=0 Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.934073 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4dbl" event={"ID":"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc","Type":"ContainerDied","Data":"20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29"} Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.934123 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4dbl" event={"ID":"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc","Type":"ContainerStarted","Data":"3802478fafcc57c03609da93256b5af6ce826db82b32b9102ad6f9714bc7ab79"} Sep 29 19:45:31 crc kubenswrapper[4741]: I0929 19:45:31.935918 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 19:45:32 crc kubenswrapper[4741]: I0929 19:45:32.946987 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4dbl" event={"ID":"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc","Type":"ContainerStarted","Data":"256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84"} Sep 29 19:45:33 crc kubenswrapper[4741]: I0929 19:45:33.955592 4741 generic.go:334] "Generic (PLEG): container finished" podID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerID="256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84" exitCode=0 Sep 29 19:45:33 crc kubenswrapper[4741]: I0929 19:45:33.955664 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4dbl" event={"ID":"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc","Type":"ContainerDied","Data":"256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84"} Sep 29 19:45:34 crc kubenswrapper[4741]: I0929 19:45:34.974210 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4dbl" event={"ID":"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc","Type":"ContainerStarted","Data":"ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302"} Sep 29 19:45:34 crc kubenswrapper[4741]: I0929 19:45:34.996357 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q4dbl" podStartSLOduration=2.585133841 podStartE2EDuration="4.996339634s" podCreationTimestamp="2025-09-29 19:45:30 +0000 UTC" firstStartedPulling="2025-09-29 19:45:31.935639777 +0000 UTC m=+2173.583429109" lastFinishedPulling="2025-09-29 19:45:34.34684556 +0000 UTC m=+2175.994634902" observedRunningTime="2025-09-29 19:45:34.990137059 +0000 UTC m=+2176.637926391" watchObservedRunningTime="2025-09-29 19:45:34.996339634 +0000 UTC m=+2176.644128956" Sep 29 19:45:41 crc kubenswrapper[4741]: I0929 19:45:41.157849 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:41 crc kubenswrapper[4741]: I0929 19:45:41.158830 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:41 crc kubenswrapper[4741]: I0929 19:45:41.210741 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:42 crc kubenswrapper[4741]: I0929 19:45:42.075223 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:42 crc kubenswrapper[4741]: I0929 19:45:42.133065 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q4dbl"] Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.046472 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q4dbl" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="registry-server" containerID="cri-o://ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302" gracePeriod=2 Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.465411 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.644617 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-catalog-content\") pod \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.644767 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qnxk\" (UniqueName: \"kubernetes.io/projected/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-kube-api-access-6qnxk\") pod \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.644949 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-utilities\") pod \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\" (UID: \"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc\") " Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.645741 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-utilities" (OuterVolumeSpecName: "utilities") pod "d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" (UID: "d2dc3ef2-2497-40f8-9558-1e3acb02b4dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.655269 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-kube-api-access-6qnxk" (OuterVolumeSpecName: "kube-api-access-6qnxk") pod "d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" (UID: "d2dc3ef2-2497-40f8-9558-1e3acb02b4dc"). InnerVolumeSpecName "kube-api-access-6qnxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.692140 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" (UID: "d2dc3ef2-2497-40f8-9558-1e3acb02b4dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.746338 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.746369 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:45:44 crc kubenswrapper[4741]: I0929 19:45:44.746379 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qnxk\" (UniqueName: \"kubernetes.io/projected/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc-kube-api-access-6qnxk\") on node \"crc\" DevicePath \"\"" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.054987 4741 generic.go:334] "Generic (PLEG): container finished" podID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerID="ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302" exitCode=0 Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.055030 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4dbl" event={"ID":"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc","Type":"ContainerDied","Data":"ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302"} Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.055059 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4dbl" event={"ID":"d2dc3ef2-2497-40f8-9558-1e3acb02b4dc","Type":"ContainerDied","Data":"3802478fafcc57c03609da93256b5af6ce826db82b32b9102ad6f9714bc7ab79"} Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.055080 4741 scope.go:117] "RemoveContainer" containerID="ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.055272 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4dbl" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.074829 4741 scope.go:117] "RemoveContainer" containerID="256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.105780 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q4dbl"] Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.113710 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q4dbl"] Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.119287 4741 scope.go:117] "RemoveContainer" containerID="20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.134509 4741 scope.go:117] "RemoveContainer" containerID="ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302" Sep 29 19:45:45 crc kubenswrapper[4741]: E0929 19:45:45.134920 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302\": container with ID starting with ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302 not found: ID does not exist" containerID="ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.134954 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302"} err="failed to get container status \"ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302\": rpc error: code = NotFound desc = could not find container \"ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302\": container with ID starting with ee2ca93b116d87474ec3eebaaea6a3abcebc79fd51df9d34dd20d4fee3224302 not found: ID does not exist" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.134976 4741 scope.go:117] "RemoveContainer" containerID="256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84" Sep 29 19:45:45 crc kubenswrapper[4741]: E0929 19:45:45.135341 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84\": container with ID starting with 256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84 not found: ID does not exist" containerID="256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.135374 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84"} err="failed to get container status \"256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84\": rpc error: code = NotFound desc = could not find container \"256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84\": container with ID starting with 256a8853e5243971f738b48132aa90705afc05833c531a2b47d4550f2dfe9c84 not found: ID does not exist" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.135415 4741 scope.go:117] "RemoveContainer" containerID="20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29" Sep 29 19:45:45 crc kubenswrapper[4741]: E0929 19:45:45.135644 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29\": container with ID starting with 20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29 not found: ID does not exist" containerID="20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29" Sep 29 19:45:45 crc kubenswrapper[4741]: I0929 19:45:45.135671 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29"} err="failed to get container status \"20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29\": rpc error: code = NotFound desc = could not find container \"20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29\": container with ID starting with 20eafa04d4b8342f03b89016744536245f407c54175caa019c039831c7c64f29 not found: ID does not exist" Sep 29 19:45:47 crc kubenswrapper[4741]: I0929 19:45:47.104206 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" path="/var/lib/kubelet/pods/d2dc3ef2-2497-40f8-9558-1e3acb02b4dc/volumes" Sep 29 19:47:31 crc kubenswrapper[4741]: I0929 19:47:31.738488 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:47:31 crc kubenswrapper[4741]: I0929 19:47:31.739237 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:48:01 crc kubenswrapper[4741]: I0929 19:48:01.738913 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:48:01 crc kubenswrapper[4741]: I0929 19:48:01.739564 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:48:31 crc kubenswrapper[4741]: I0929 19:48:31.738830 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:48:31 crc kubenswrapper[4741]: I0929 19:48:31.739310 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:48:31 crc kubenswrapper[4741]: I0929 19:48:31.739354 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:48:31 crc kubenswrapper[4741]: I0929 19:48:31.740013 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:48:31 crc kubenswrapper[4741]: I0929 19:48:31.740069 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" gracePeriod=600 Sep 29 19:48:31 crc kubenswrapper[4741]: E0929 19:48:31.866263 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:48:32 crc kubenswrapper[4741]: I0929 19:48:32.381693 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" exitCode=0 Sep 29 19:48:32 crc kubenswrapper[4741]: I0929 19:48:32.381755 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff"} Sep 29 19:48:32 crc kubenswrapper[4741]: I0929 19:48:32.381864 4741 scope.go:117] "RemoveContainer" containerID="3a8181336d2441af20e9c2452f4c8711ec838be48460263fc0c16a5200c4fdcf" Sep 29 19:48:32 crc kubenswrapper[4741]: I0929 19:48:32.382594 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:48:32 crc kubenswrapper[4741]: E0929 19:48:32.382993 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:48:47 crc kubenswrapper[4741]: I0929 19:48:47.086424 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:48:47 crc kubenswrapper[4741]: E0929 19:48:47.087297 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:48:58 crc kubenswrapper[4741]: I0929 19:48:58.086319 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:48:58 crc kubenswrapper[4741]: E0929 19:48:58.087122 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:49:12 crc kubenswrapper[4741]: I0929 19:49:12.085848 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:49:12 crc kubenswrapper[4741]: E0929 19:49:12.086681 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:49:25 crc kubenswrapper[4741]: I0929 19:49:25.086898 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:49:25 crc kubenswrapper[4741]: E0929 19:49:25.089503 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:49:40 crc kubenswrapper[4741]: I0929 19:49:40.086197 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:49:40 crc kubenswrapper[4741]: E0929 19:49:40.086904 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:49:53 crc kubenswrapper[4741]: I0929 19:49:53.086334 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:49:53 crc kubenswrapper[4741]: E0929 19:49:53.087326 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:50:07 crc kubenswrapper[4741]: I0929 19:50:07.086249 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:50:07 crc kubenswrapper[4741]: E0929 19:50:07.086933 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.086355 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:50:20 crc kubenswrapper[4741]: E0929 19:50:20.087362 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.746497 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kfjlm"] Sep 29 19:50:20 crc kubenswrapper[4741]: E0929 19:50:20.747093 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="extract-utilities" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.747130 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="extract-utilities" Sep 29 19:50:20 crc kubenswrapper[4741]: E0929 19:50:20.747165 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="registry-server" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.747184 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="registry-server" Sep 29 19:50:20 crc kubenswrapper[4741]: E0929 19:50:20.747223 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="extract-content" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.747238 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="extract-content" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.747570 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2dc3ef2-2497-40f8-9558-1e3acb02b4dc" containerName="registry-server" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.750145 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.768677 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfjlm"] Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.899145 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-utilities\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.899240 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-catalog-content\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:20 crc kubenswrapper[4741]: I0929 19:50:20.899283 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq8vv\" (UniqueName: \"kubernetes.io/projected/50a5a153-a107-4391-858d-fc4ffa2d656b-kube-api-access-sq8vv\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.000563 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-catalog-content\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.001201 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-catalog-content\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.001491 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq8vv\" (UniqueName: \"kubernetes.io/projected/50a5a153-a107-4391-858d-fc4ffa2d656b-kube-api-access-sq8vv\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.001748 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-utilities\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.002250 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-utilities\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.026752 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq8vv\" (UniqueName: \"kubernetes.io/projected/50a5a153-a107-4391-858d-fc4ffa2d656b-kube-api-access-sq8vv\") pod \"certified-operators-kfjlm\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.081415 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:21 crc kubenswrapper[4741]: I0929 19:50:21.339763 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfjlm"] Sep 29 19:50:22 crc kubenswrapper[4741]: I0929 19:50:22.294545 4741 generic.go:334] "Generic (PLEG): container finished" podID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerID="41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762" exitCode=0 Sep 29 19:50:22 crc kubenswrapper[4741]: I0929 19:50:22.294626 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfjlm" event={"ID":"50a5a153-a107-4391-858d-fc4ffa2d656b","Type":"ContainerDied","Data":"41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762"} Sep 29 19:50:22 crc kubenswrapper[4741]: I0929 19:50:22.295104 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfjlm" event={"ID":"50a5a153-a107-4391-858d-fc4ffa2d656b","Type":"ContainerStarted","Data":"15b2971b555980ae72e15027acd133417ca8022997b319c19f8f7ea2dd199112"} Sep 29 19:50:23 crc kubenswrapper[4741]: I0929 19:50:23.303859 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfjlm" event={"ID":"50a5a153-a107-4391-858d-fc4ffa2d656b","Type":"ContainerStarted","Data":"ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738"} Sep 29 19:50:24 crc kubenswrapper[4741]: I0929 19:50:24.315763 4741 generic.go:334] "Generic (PLEG): container finished" podID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerID="ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738" exitCode=0 Sep 29 19:50:24 crc kubenswrapper[4741]: I0929 19:50:24.315850 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfjlm" event={"ID":"50a5a153-a107-4391-858d-fc4ffa2d656b","Type":"ContainerDied","Data":"ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738"} Sep 29 19:50:25 crc kubenswrapper[4741]: I0929 19:50:25.326981 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfjlm" event={"ID":"50a5a153-a107-4391-858d-fc4ffa2d656b","Type":"ContainerStarted","Data":"fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043"} Sep 29 19:50:25 crc kubenswrapper[4741]: I0929 19:50:25.354045 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kfjlm" podStartSLOduration=2.913064481 podStartE2EDuration="5.354026302s" podCreationTimestamp="2025-09-29 19:50:20 +0000 UTC" firstStartedPulling="2025-09-29 19:50:22.297270565 +0000 UTC m=+2463.945059897" lastFinishedPulling="2025-09-29 19:50:24.738232396 +0000 UTC m=+2466.386021718" observedRunningTime="2025-09-29 19:50:25.345589646 +0000 UTC m=+2466.993378998" watchObservedRunningTime="2025-09-29 19:50:25.354026302 +0000 UTC m=+2467.001815634" Sep 29 19:50:31 crc kubenswrapper[4741]: I0929 19:50:31.084070 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:31 crc kubenswrapper[4741]: I0929 19:50:31.084837 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:31 crc kubenswrapper[4741]: I0929 19:50:31.146926 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:31 crc kubenswrapper[4741]: I0929 19:50:31.415851 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:31 crc kubenswrapper[4741]: I0929 19:50:31.465687 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfjlm"] Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.086208 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:50:33 crc kubenswrapper[4741]: E0929 19:50:33.086982 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.389688 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kfjlm" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="registry-server" containerID="cri-o://fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043" gracePeriod=2 Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.812103 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.991970 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-catalog-content\") pod \"50a5a153-a107-4391-858d-fc4ffa2d656b\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.992050 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq8vv\" (UniqueName: \"kubernetes.io/projected/50a5a153-a107-4391-858d-fc4ffa2d656b-kube-api-access-sq8vv\") pod \"50a5a153-a107-4391-858d-fc4ffa2d656b\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.992155 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-utilities\") pod \"50a5a153-a107-4391-858d-fc4ffa2d656b\" (UID: \"50a5a153-a107-4391-858d-fc4ffa2d656b\") " Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.993026 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-utilities" (OuterVolumeSpecName: "utilities") pod "50a5a153-a107-4391-858d-fc4ffa2d656b" (UID: "50a5a153-a107-4391-858d-fc4ffa2d656b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:50:33 crc kubenswrapper[4741]: I0929 19:50:33.997037 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a5a153-a107-4391-858d-fc4ffa2d656b-kube-api-access-sq8vv" (OuterVolumeSpecName: "kube-api-access-sq8vv") pod "50a5a153-a107-4391-858d-fc4ffa2d656b" (UID: "50a5a153-a107-4391-858d-fc4ffa2d656b"). InnerVolumeSpecName "kube-api-access-sq8vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.043938 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50a5a153-a107-4391-858d-fc4ffa2d656b" (UID: "50a5a153-a107-4391-858d-fc4ffa2d656b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.094000 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.094025 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a5a153-a107-4391-858d-fc4ffa2d656b-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.094038 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq8vv\" (UniqueName: \"kubernetes.io/projected/50a5a153-a107-4391-858d-fc4ffa2d656b-kube-api-access-sq8vv\") on node \"crc\" DevicePath \"\"" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.403840 4741 generic.go:334] "Generic (PLEG): container finished" podID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerID="fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043" exitCode=0 Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.403882 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfjlm" event={"ID":"50a5a153-a107-4391-858d-fc4ffa2d656b","Type":"ContainerDied","Data":"fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043"} Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.403910 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfjlm" event={"ID":"50a5a153-a107-4391-858d-fc4ffa2d656b","Type":"ContainerDied","Data":"15b2971b555980ae72e15027acd133417ca8022997b319c19f8f7ea2dd199112"} Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.403929 4741 scope.go:117] "RemoveContainer" containerID="fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.404326 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfjlm" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.429286 4741 scope.go:117] "RemoveContainer" containerID="ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.456120 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfjlm"] Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.460734 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kfjlm"] Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.468504 4741 scope.go:117] "RemoveContainer" containerID="41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.508658 4741 scope.go:117] "RemoveContainer" containerID="fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043" Sep 29 19:50:34 crc kubenswrapper[4741]: E0929 19:50:34.509178 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043\": container with ID starting with fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043 not found: ID does not exist" containerID="fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.509222 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043"} err="failed to get container status \"fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043\": rpc error: code = NotFound desc = could not find container \"fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043\": container with ID starting with fa906d873e43c7429035016aaa10bca67ea36d3f2039d10904d850f5498f6043 not found: ID does not exist" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.509245 4741 scope.go:117] "RemoveContainer" containerID="ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738" Sep 29 19:50:34 crc kubenswrapper[4741]: E0929 19:50:34.509616 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738\": container with ID starting with ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738 not found: ID does not exist" containerID="ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.509724 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738"} err="failed to get container status \"ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738\": rpc error: code = NotFound desc = could not find container \"ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738\": container with ID starting with ecee6fce855aa8be7bedb394ef38fbd069fa1e74278568115d3fed372a319738 not found: ID does not exist" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.509798 4741 scope.go:117] "RemoveContainer" containerID="41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762" Sep 29 19:50:34 crc kubenswrapper[4741]: E0929 19:50:34.510095 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762\": container with ID starting with 41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762 not found: ID does not exist" containerID="41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762" Sep 29 19:50:34 crc kubenswrapper[4741]: I0929 19:50:34.510126 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762"} err="failed to get container status \"41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762\": rpc error: code = NotFound desc = could not find container \"41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762\": container with ID starting with 41c9aee65ac45280c1f36dacd3e1a895c9e6ec6c7ad570ece2c70f156976e762 not found: ID does not exist" Sep 29 19:50:35 crc kubenswrapper[4741]: I0929 19:50:35.101565 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" path="/var/lib/kubelet/pods/50a5a153-a107-4391-858d-fc4ffa2d656b/volumes" Sep 29 19:50:46 crc kubenswrapper[4741]: I0929 19:50:46.086523 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:50:46 crc kubenswrapper[4741]: E0929 19:50:46.087423 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:51:00 crc kubenswrapper[4741]: I0929 19:51:00.087098 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:51:00 crc kubenswrapper[4741]: E0929 19:51:00.088128 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:51:15 crc kubenswrapper[4741]: I0929 19:51:15.086457 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:51:15 crc kubenswrapper[4741]: E0929 19:51:15.087206 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:51:29 crc kubenswrapper[4741]: I0929 19:51:29.091690 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:51:29 crc kubenswrapper[4741]: E0929 19:51:29.094899 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:51:44 crc kubenswrapper[4741]: I0929 19:51:44.086307 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:51:44 crc kubenswrapper[4741]: E0929 19:51:44.086975 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:51:58 crc kubenswrapper[4741]: I0929 19:51:58.086006 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:51:58 crc kubenswrapper[4741]: E0929 19:51:58.087244 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:52:11 crc kubenswrapper[4741]: I0929 19:52:11.086313 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:52:11 crc kubenswrapper[4741]: E0929 19:52:11.089031 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:52:23 crc kubenswrapper[4741]: I0929 19:52:23.086079 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:52:23 crc kubenswrapper[4741]: E0929 19:52:23.086792 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:52:34 crc kubenswrapper[4741]: I0929 19:52:34.085963 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:52:34 crc kubenswrapper[4741]: E0929 19:52:34.086739 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:52:49 crc kubenswrapper[4741]: I0929 19:52:49.089637 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:52:49 crc kubenswrapper[4741]: E0929 19:52:49.090554 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:53:02 crc kubenswrapper[4741]: I0929 19:53:02.086208 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:53:02 crc kubenswrapper[4741]: E0929 19:53:02.086910 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:53:15 crc kubenswrapper[4741]: I0929 19:53:15.086458 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:53:15 crc kubenswrapper[4741]: E0929 19:53:15.089100 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.512889 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-csg22"] Sep 29 19:53:19 crc kubenswrapper[4741]: E0929 19:53:19.514038 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="registry-server" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.514057 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="registry-server" Sep 29 19:53:19 crc kubenswrapper[4741]: E0929 19:53:19.514070 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="extract-utilities" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.514080 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="extract-utilities" Sep 29 19:53:19 crc kubenswrapper[4741]: E0929 19:53:19.514103 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="extract-content" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.514113 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="extract-content" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.514357 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a5a153-a107-4391-858d-fc4ffa2d656b" containerName="registry-server" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.515971 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.535604 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-csg22"] Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.547226 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-utilities\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.547329 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmp59\" (UniqueName: \"kubernetes.io/projected/a5f04ca9-6f68-43ba-b3e2-33416bff266d-kube-api-access-bmp59\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.547369 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-catalog-content\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.648304 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-utilities\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.648379 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmp59\" (UniqueName: \"kubernetes.io/projected/a5f04ca9-6f68-43ba-b3e2-33416bff266d-kube-api-access-bmp59\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.648421 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-catalog-content\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.648960 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-catalog-content\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.649112 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-utilities\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.673525 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmp59\" (UniqueName: \"kubernetes.io/projected/a5f04ca9-6f68-43ba-b3e2-33416bff266d-kube-api-access-bmp59\") pod \"redhat-marketplace-csg22\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:19 crc kubenswrapper[4741]: I0929 19:53:19.844171 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:20 crc kubenswrapper[4741]: I0929 19:53:20.049831 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-csg22"] Sep 29 19:53:20 crc kubenswrapper[4741]: I0929 19:53:20.778880 4741 generic.go:334] "Generic (PLEG): container finished" podID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerID="c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5" exitCode=0 Sep 29 19:53:20 crc kubenswrapper[4741]: I0929 19:53:20.778929 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csg22" event={"ID":"a5f04ca9-6f68-43ba-b3e2-33416bff266d","Type":"ContainerDied","Data":"c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5"} Sep 29 19:53:20 crc kubenswrapper[4741]: I0929 19:53:20.778954 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csg22" event={"ID":"a5f04ca9-6f68-43ba-b3e2-33416bff266d","Type":"ContainerStarted","Data":"83c99217b339f4555d3e20645305d9fe3a65c3af0c81aa1fdd4ffb2030513fd3"} Sep 29 19:53:20 crc kubenswrapper[4741]: I0929 19:53:20.782490 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 19:53:21 crc kubenswrapper[4741]: I0929 19:53:21.789006 4741 generic.go:334] "Generic (PLEG): container finished" podID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerID="30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73" exitCode=0 Sep 29 19:53:21 crc kubenswrapper[4741]: I0929 19:53:21.789118 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csg22" event={"ID":"a5f04ca9-6f68-43ba-b3e2-33416bff266d","Type":"ContainerDied","Data":"30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73"} Sep 29 19:53:22 crc kubenswrapper[4741]: I0929 19:53:22.796812 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csg22" event={"ID":"a5f04ca9-6f68-43ba-b3e2-33416bff266d","Type":"ContainerStarted","Data":"7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b"} Sep 29 19:53:22 crc kubenswrapper[4741]: I0929 19:53:22.815599 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-csg22" podStartSLOduration=2.348493838 podStartE2EDuration="3.815579415s" podCreationTimestamp="2025-09-29 19:53:19 +0000 UTC" firstStartedPulling="2025-09-29 19:53:20.781213875 +0000 UTC m=+2642.429003217" lastFinishedPulling="2025-09-29 19:53:22.248299442 +0000 UTC m=+2643.896088794" observedRunningTime="2025-09-29 19:53:22.813199169 +0000 UTC m=+2644.460988501" watchObservedRunningTime="2025-09-29 19:53:22.815579415 +0000 UTC m=+2644.463368747" Sep 29 19:53:27 crc kubenswrapper[4741]: I0929 19:53:27.086420 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:53:27 crc kubenswrapper[4741]: E0929 19:53:27.086946 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 19:53:29 crc kubenswrapper[4741]: I0929 19:53:29.844887 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:29 crc kubenswrapper[4741]: I0929 19:53:29.845220 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:29 crc kubenswrapper[4741]: I0929 19:53:29.929185 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:29 crc kubenswrapper[4741]: I0929 19:53:29.997459 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:30 crc kubenswrapper[4741]: I0929 19:53:30.170552 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-csg22"] Sep 29 19:53:31 crc kubenswrapper[4741]: I0929 19:53:31.874582 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-csg22" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="registry-server" containerID="cri-o://7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b" gracePeriod=2 Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.230364 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.259082 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-utilities\") pod \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.259157 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-catalog-content\") pod \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.259178 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmp59\" (UniqueName: \"kubernetes.io/projected/a5f04ca9-6f68-43ba-b3e2-33416bff266d-kube-api-access-bmp59\") pod \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\" (UID: \"a5f04ca9-6f68-43ba-b3e2-33416bff266d\") " Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.261332 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-utilities" (OuterVolumeSpecName: "utilities") pod "a5f04ca9-6f68-43ba-b3e2-33416bff266d" (UID: "a5f04ca9-6f68-43ba-b3e2-33416bff266d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.264870 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f04ca9-6f68-43ba-b3e2-33416bff266d-kube-api-access-bmp59" (OuterVolumeSpecName: "kube-api-access-bmp59") pod "a5f04ca9-6f68-43ba-b3e2-33416bff266d" (UID: "a5f04ca9-6f68-43ba-b3e2-33416bff266d"). InnerVolumeSpecName "kube-api-access-bmp59". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.273757 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5f04ca9-6f68-43ba-b3e2-33416bff266d" (UID: "a5f04ca9-6f68-43ba-b3e2-33416bff266d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.360546 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.360574 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f04ca9-6f68-43ba-b3e2-33416bff266d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.360584 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmp59\" (UniqueName: \"kubernetes.io/projected/a5f04ca9-6f68-43ba-b3e2-33416bff266d-kube-api-access-bmp59\") on node \"crc\" DevicePath \"\"" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.882744 4741 generic.go:334] "Generic (PLEG): container finished" podID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerID="7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b" exitCode=0 Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.882813 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csg22" event={"ID":"a5f04ca9-6f68-43ba-b3e2-33416bff266d","Type":"ContainerDied","Data":"7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b"} Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.882852 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csg22" event={"ID":"a5f04ca9-6f68-43ba-b3e2-33416bff266d","Type":"ContainerDied","Data":"83c99217b339f4555d3e20645305d9fe3a65c3af0c81aa1fdd4ffb2030513fd3"} Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.882880 4741 scope.go:117] "RemoveContainer" containerID="7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.883066 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csg22" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.899656 4741 scope.go:117] "RemoveContainer" containerID="30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.921769 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-csg22"] Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.929591 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-csg22"] Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.936451 4741 scope.go:117] "RemoveContainer" containerID="c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.952129 4741 scope.go:117] "RemoveContainer" containerID="7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b" Sep 29 19:53:32 crc kubenswrapper[4741]: E0929 19:53:32.952535 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b\": container with ID starting with 7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b not found: ID does not exist" containerID="7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.952571 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b"} err="failed to get container status \"7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b\": rpc error: code = NotFound desc = could not find container \"7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b\": container with ID starting with 7d491477ff44b698626a751738d2e96baff7508e9120d211d8a30230a05f271b not found: ID does not exist" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.952598 4741 scope.go:117] "RemoveContainer" containerID="30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73" Sep 29 19:53:32 crc kubenswrapper[4741]: E0929 19:53:32.952948 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73\": container with ID starting with 30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73 not found: ID does not exist" containerID="30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.952976 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73"} err="failed to get container status \"30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73\": rpc error: code = NotFound desc = could not find container \"30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73\": container with ID starting with 30cdb20f40f35f6a87b341064522f0e963f92a50794b613d2b6f58c796d2fd73 not found: ID does not exist" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.952994 4741 scope.go:117] "RemoveContainer" containerID="c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5" Sep 29 19:53:32 crc kubenswrapper[4741]: E0929 19:53:32.953231 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5\": container with ID starting with c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5 not found: ID does not exist" containerID="c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5" Sep 29 19:53:32 crc kubenswrapper[4741]: I0929 19:53:32.953257 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5"} err="failed to get container status \"c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5\": rpc error: code = NotFound desc = could not find container \"c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5\": container with ID starting with c98416169f67729765e60857f4ba373f075d5c23009d5418f39f3883b4adf3e5 not found: ID does not exist" Sep 29 19:53:33 crc kubenswrapper[4741]: I0929 19:53:33.094725 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" path="/var/lib/kubelet/pods/a5f04ca9-6f68-43ba-b3e2-33416bff266d/volumes" Sep 29 19:53:40 crc kubenswrapper[4741]: I0929 19:53:40.085419 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:53:40 crc kubenswrapper[4741]: I0929 19:53:40.959410 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"89ba566aa7ea8f6245d4f453a81ad158f7f9415979df78f7544100d5299c5ac0"} Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.719475 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6cvwc"] Sep 29 19:53:43 crc kubenswrapper[4741]: E0929 19:53:43.720319 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="extract-content" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.720331 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="extract-content" Sep 29 19:53:43 crc kubenswrapper[4741]: E0929 19:53:43.720349 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="registry-server" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.720356 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="registry-server" Sep 29 19:53:43 crc kubenswrapper[4741]: E0929 19:53:43.720377 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="extract-utilities" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.720383 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="extract-utilities" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.720578 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f04ca9-6f68-43ba-b3e2-33416bff266d" containerName="registry-server" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.721811 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.734501 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6cvwc"] Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.828312 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-catalog-content\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.828367 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjxcl\" (UniqueName: \"kubernetes.io/projected/09f1891b-87b8-4563-9814-d130ec3ba525-kube-api-access-qjxcl\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.828635 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-utilities\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.930005 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-utilities\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.930082 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-catalog-content\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.930112 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjxcl\" (UniqueName: \"kubernetes.io/projected/09f1891b-87b8-4563-9814-d130ec3ba525-kube-api-access-qjxcl\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.930644 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-utilities\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.930658 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-catalog-content\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:43 crc kubenswrapper[4741]: I0929 19:53:43.954772 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjxcl\" (UniqueName: \"kubernetes.io/projected/09f1891b-87b8-4563-9814-d130ec3ba525-kube-api-access-qjxcl\") pod \"redhat-operators-6cvwc\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:44 crc kubenswrapper[4741]: I0929 19:53:44.038888 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:44 crc kubenswrapper[4741]: I0929 19:53:44.268165 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6cvwc"] Sep 29 19:53:44 crc kubenswrapper[4741]: I0929 19:53:44.986673 4741 generic.go:334] "Generic (PLEG): container finished" podID="09f1891b-87b8-4563-9814-d130ec3ba525" containerID="882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc" exitCode=0 Sep 29 19:53:44 crc kubenswrapper[4741]: I0929 19:53:44.986729 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cvwc" event={"ID":"09f1891b-87b8-4563-9814-d130ec3ba525","Type":"ContainerDied","Data":"882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc"} Sep 29 19:53:44 crc kubenswrapper[4741]: I0929 19:53:44.987252 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cvwc" event={"ID":"09f1891b-87b8-4563-9814-d130ec3ba525","Type":"ContainerStarted","Data":"f945d462293c5f09a622572eb2adfbfc7b28cecdd77e42a3bca1605c500f18b9"} Sep 29 19:53:47 crc kubenswrapper[4741]: I0929 19:53:47.003361 4741 generic.go:334] "Generic (PLEG): container finished" podID="09f1891b-87b8-4563-9814-d130ec3ba525" containerID="feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c" exitCode=0 Sep 29 19:53:47 crc kubenswrapper[4741]: I0929 19:53:47.003432 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cvwc" event={"ID":"09f1891b-87b8-4563-9814-d130ec3ba525","Type":"ContainerDied","Data":"feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c"} Sep 29 19:53:48 crc kubenswrapper[4741]: I0929 19:53:48.012943 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cvwc" event={"ID":"09f1891b-87b8-4563-9814-d130ec3ba525","Type":"ContainerStarted","Data":"a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016"} Sep 29 19:53:48 crc kubenswrapper[4741]: I0929 19:53:48.033633 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6cvwc" podStartSLOduration=2.627256412 podStartE2EDuration="5.033615831s" podCreationTimestamp="2025-09-29 19:53:43 +0000 UTC" firstStartedPulling="2025-09-29 19:53:44.988434808 +0000 UTC m=+2666.636224140" lastFinishedPulling="2025-09-29 19:53:47.394794217 +0000 UTC m=+2669.042583559" observedRunningTime="2025-09-29 19:53:48.029114508 +0000 UTC m=+2669.676903840" watchObservedRunningTime="2025-09-29 19:53:48.033615831 +0000 UTC m=+2669.681405153" Sep 29 19:53:54 crc kubenswrapper[4741]: I0929 19:53:54.039621 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:54 crc kubenswrapper[4741]: I0929 19:53:54.040463 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:54 crc kubenswrapper[4741]: I0929 19:53:54.097158 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:54 crc kubenswrapper[4741]: I0929 19:53:54.152159 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:55 crc kubenswrapper[4741]: I0929 19:53:55.709485 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6cvwc"] Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.069241 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6cvwc" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="registry-server" containerID="cri-o://a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016" gracePeriod=2 Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.424178 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.590829 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjxcl\" (UniqueName: \"kubernetes.io/projected/09f1891b-87b8-4563-9814-d130ec3ba525-kube-api-access-qjxcl\") pod \"09f1891b-87b8-4563-9814-d130ec3ba525\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.590881 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-utilities\") pod \"09f1891b-87b8-4563-9814-d130ec3ba525\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.590932 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-catalog-content\") pod \"09f1891b-87b8-4563-9814-d130ec3ba525\" (UID: \"09f1891b-87b8-4563-9814-d130ec3ba525\") " Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.592009 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-utilities" (OuterVolumeSpecName: "utilities") pod "09f1891b-87b8-4563-9814-d130ec3ba525" (UID: "09f1891b-87b8-4563-9814-d130ec3ba525"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.611292 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f1891b-87b8-4563-9814-d130ec3ba525-kube-api-access-qjxcl" (OuterVolumeSpecName: "kube-api-access-qjxcl") pod "09f1891b-87b8-4563-9814-d130ec3ba525" (UID: "09f1891b-87b8-4563-9814-d130ec3ba525"). InnerVolumeSpecName "kube-api-access-qjxcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.692138 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjxcl\" (UniqueName: \"kubernetes.io/projected/09f1891b-87b8-4563-9814-d130ec3ba525-kube-api-access-qjxcl\") on node \"crc\" DevicePath \"\"" Sep 29 19:53:56 crc kubenswrapper[4741]: I0929 19:53:56.692186 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.077837 4741 generic.go:334] "Generic (PLEG): container finished" podID="09f1891b-87b8-4563-9814-d130ec3ba525" containerID="a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016" exitCode=0 Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.077878 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cvwc" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.077881 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cvwc" event={"ID":"09f1891b-87b8-4563-9814-d130ec3ba525","Type":"ContainerDied","Data":"a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016"} Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.078003 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cvwc" event={"ID":"09f1891b-87b8-4563-9814-d130ec3ba525","Type":"ContainerDied","Data":"f945d462293c5f09a622572eb2adfbfc7b28cecdd77e42a3bca1605c500f18b9"} Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.078029 4741 scope.go:117] "RemoveContainer" containerID="a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.095557 4741 scope.go:117] "RemoveContainer" containerID="feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.114955 4741 scope.go:117] "RemoveContainer" containerID="882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.137000 4741 scope.go:117] "RemoveContainer" containerID="a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016" Sep 29 19:53:57 crc kubenswrapper[4741]: E0929 19:53:57.137441 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016\": container with ID starting with a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016 not found: ID does not exist" containerID="a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.137485 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016"} err="failed to get container status \"a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016\": rpc error: code = NotFound desc = could not find container \"a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016\": container with ID starting with a7cfbc26c9537085ad1d8896e8ca64eebc4822a949f1dfee5705128b8eda4016 not found: ID does not exist" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.137513 4741 scope.go:117] "RemoveContainer" containerID="feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c" Sep 29 19:53:57 crc kubenswrapper[4741]: E0929 19:53:57.137789 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c\": container with ID starting with feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c not found: ID does not exist" containerID="feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.137910 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c"} err="failed to get container status \"feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c\": rpc error: code = NotFound desc = could not find container \"feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c\": container with ID starting with feb15319b751d658f0bc35d91e395ef5fa7c4eeabcf4a37c72119a1ebbecaf9c not found: ID does not exist" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.137998 4741 scope.go:117] "RemoveContainer" containerID="882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc" Sep 29 19:53:57 crc kubenswrapper[4741]: E0929 19:53:57.138287 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc\": container with ID starting with 882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc not found: ID does not exist" containerID="882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.138375 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc"} err="failed to get container status \"882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc\": rpc error: code = NotFound desc = could not find container \"882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc\": container with ID starting with 882483aafa78a9f647368fc6daa3d76559f47a1469b2e9a6c42f6bfda2c8c4cc not found: ID does not exist" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.664581 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09f1891b-87b8-4563-9814-d130ec3ba525" (UID: "09f1891b-87b8-4563-9814-d130ec3ba525"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.705462 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f1891b-87b8-4563-9814-d130ec3ba525-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.718709 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6cvwc"] Sep 29 19:53:57 crc kubenswrapper[4741]: I0929 19:53:57.724692 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6cvwc"] Sep 29 19:53:59 crc kubenswrapper[4741]: I0929 19:53:59.093866 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" path="/var/lib/kubelet/pods/09f1891b-87b8-4563-9814-d130ec3ba525/volumes" Sep 29 19:56:01 crc kubenswrapper[4741]: I0929 19:56:01.740149 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:56:01 crc kubenswrapper[4741]: I0929 19:56:01.740898 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.535069 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sjvkz"] Sep 29 19:56:04 crc kubenswrapper[4741]: E0929 19:56:04.536098 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="extract-content" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.536122 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="extract-content" Sep 29 19:56:04 crc kubenswrapper[4741]: E0929 19:56:04.536139 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="registry-server" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.536179 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="registry-server" Sep 29 19:56:04 crc kubenswrapper[4741]: E0929 19:56:04.536219 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="extract-utilities" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.536232 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="extract-utilities" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.536705 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f1891b-87b8-4563-9814-d130ec3ba525" containerName="registry-server" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.538682 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.549465 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sjvkz"] Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.649130 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-utilities\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.649364 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj9qq\" (UniqueName: \"kubernetes.io/projected/6983b04b-ca22-4509-ad2d-c8d067a2709f-kube-api-access-mj9qq\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.649510 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-catalog-content\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.751252 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-utilities\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.751327 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj9qq\" (UniqueName: \"kubernetes.io/projected/6983b04b-ca22-4509-ad2d-c8d067a2709f-kube-api-access-mj9qq\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.751359 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-catalog-content\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.751891 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-catalog-content\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.751894 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-utilities\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.774295 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj9qq\" (UniqueName: \"kubernetes.io/projected/6983b04b-ca22-4509-ad2d-c8d067a2709f-kube-api-access-mj9qq\") pod \"community-operators-sjvkz\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:04 crc kubenswrapper[4741]: I0929 19:56:04.898554 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:05 crc kubenswrapper[4741]: I0929 19:56:05.354777 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sjvkz"] Sep 29 19:56:06 crc kubenswrapper[4741]: I0929 19:56:06.130771 4741 generic.go:334] "Generic (PLEG): container finished" podID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerID="98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9" exitCode=0 Sep 29 19:56:06 crc kubenswrapper[4741]: I0929 19:56:06.130809 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjvkz" event={"ID":"6983b04b-ca22-4509-ad2d-c8d067a2709f","Type":"ContainerDied","Data":"98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9"} Sep 29 19:56:06 crc kubenswrapper[4741]: I0929 19:56:06.132486 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjvkz" event={"ID":"6983b04b-ca22-4509-ad2d-c8d067a2709f","Type":"ContainerStarted","Data":"4522cf4b30407a9c5c6002896eaf3f0b254c78b8c391ea344170441aafb9f1dd"} Sep 29 19:56:07 crc kubenswrapper[4741]: I0929 19:56:07.142612 4741 generic.go:334] "Generic (PLEG): container finished" podID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerID="e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51" exitCode=0 Sep 29 19:56:07 crc kubenswrapper[4741]: I0929 19:56:07.142703 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjvkz" event={"ID":"6983b04b-ca22-4509-ad2d-c8d067a2709f","Type":"ContainerDied","Data":"e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51"} Sep 29 19:56:08 crc kubenswrapper[4741]: I0929 19:56:08.152267 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjvkz" event={"ID":"6983b04b-ca22-4509-ad2d-c8d067a2709f","Type":"ContainerStarted","Data":"2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f"} Sep 29 19:56:08 crc kubenswrapper[4741]: I0929 19:56:08.174446 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sjvkz" podStartSLOduration=2.620870285 podStartE2EDuration="4.17442993s" podCreationTimestamp="2025-09-29 19:56:04 +0000 UTC" firstStartedPulling="2025-09-29 19:56:06.132355906 +0000 UTC m=+2807.780145238" lastFinishedPulling="2025-09-29 19:56:07.685915551 +0000 UTC m=+2809.333704883" observedRunningTime="2025-09-29 19:56:08.17002228 +0000 UTC m=+2809.817811622" watchObservedRunningTime="2025-09-29 19:56:08.17442993 +0000 UTC m=+2809.822219262" Sep 29 19:56:14 crc kubenswrapper[4741]: I0929 19:56:14.899307 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:14 crc kubenswrapper[4741]: I0929 19:56:14.900151 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:14 crc kubenswrapper[4741]: I0929 19:56:14.990301 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:15 crc kubenswrapper[4741]: I0929 19:56:15.274499 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:15 crc kubenswrapper[4741]: I0929 19:56:15.328770 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sjvkz"] Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.226069 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sjvkz" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="registry-server" containerID="cri-o://2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f" gracePeriod=2 Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.695171 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.750405 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj9qq\" (UniqueName: \"kubernetes.io/projected/6983b04b-ca22-4509-ad2d-c8d067a2709f-kube-api-access-mj9qq\") pod \"6983b04b-ca22-4509-ad2d-c8d067a2709f\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.750458 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-catalog-content\") pod \"6983b04b-ca22-4509-ad2d-c8d067a2709f\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.750535 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-utilities\") pod \"6983b04b-ca22-4509-ad2d-c8d067a2709f\" (UID: \"6983b04b-ca22-4509-ad2d-c8d067a2709f\") " Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.751480 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-utilities" (OuterVolumeSpecName: "utilities") pod "6983b04b-ca22-4509-ad2d-c8d067a2709f" (UID: "6983b04b-ca22-4509-ad2d-c8d067a2709f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.760027 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6983b04b-ca22-4509-ad2d-c8d067a2709f-kube-api-access-mj9qq" (OuterVolumeSpecName: "kube-api-access-mj9qq") pod "6983b04b-ca22-4509-ad2d-c8d067a2709f" (UID: "6983b04b-ca22-4509-ad2d-c8d067a2709f"). InnerVolumeSpecName "kube-api-access-mj9qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.795289 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6983b04b-ca22-4509-ad2d-c8d067a2709f" (UID: "6983b04b-ca22-4509-ad2d-c8d067a2709f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.852063 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.852094 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj9qq\" (UniqueName: \"kubernetes.io/projected/6983b04b-ca22-4509-ad2d-c8d067a2709f-kube-api-access-mj9qq\") on node \"crc\" DevicePath \"\"" Sep 29 19:56:17 crc kubenswrapper[4741]: I0929 19:56:17.852107 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6983b04b-ca22-4509-ad2d-c8d067a2709f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.249255 4741 generic.go:334] "Generic (PLEG): container finished" podID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerID="2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f" exitCode=0 Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.249318 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjvkz" event={"ID":"6983b04b-ca22-4509-ad2d-c8d067a2709f","Type":"ContainerDied","Data":"2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f"} Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.249354 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjvkz" event={"ID":"6983b04b-ca22-4509-ad2d-c8d067a2709f","Type":"ContainerDied","Data":"4522cf4b30407a9c5c6002896eaf3f0b254c78b8c391ea344170441aafb9f1dd"} Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.249419 4741 scope.go:117] "RemoveContainer" containerID="2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.249848 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjvkz" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.281088 4741 scope.go:117] "RemoveContainer" containerID="e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.308524 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sjvkz"] Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.311498 4741 scope.go:117] "RemoveContainer" containerID="98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.314710 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sjvkz"] Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.336922 4741 scope.go:117] "RemoveContainer" containerID="2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f" Sep 29 19:56:18 crc kubenswrapper[4741]: E0929 19:56:18.337410 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f\": container with ID starting with 2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f not found: ID does not exist" containerID="2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.337455 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f"} err="failed to get container status \"2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f\": rpc error: code = NotFound desc = could not find container \"2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f\": container with ID starting with 2a44841a149913c15530dd35d8ffa643308ca56ecec03f9debdea9e231df320f not found: ID does not exist" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.337487 4741 scope.go:117] "RemoveContainer" containerID="e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51" Sep 29 19:56:18 crc kubenswrapper[4741]: E0929 19:56:18.337834 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51\": container with ID starting with e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51 not found: ID does not exist" containerID="e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.337882 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51"} err="failed to get container status \"e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51\": rpc error: code = NotFound desc = could not find container \"e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51\": container with ID starting with e962bee736b29d281413951c488e67cd6cacf4d210bf314484e6e5d606e8fd51 not found: ID does not exist" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.337915 4741 scope.go:117] "RemoveContainer" containerID="98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9" Sep 29 19:56:18 crc kubenswrapper[4741]: E0929 19:56:18.338315 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9\": container with ID starting with 98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9 not found: ID does not exist" containerID="98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9" Sep 29 19:56:18 crc kubenswrapper[4741]: I0929 19:56:18.338346 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9"} err="failed to get container status \"98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9\": rpc error: code = NotFound desc = could not find container \"98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9\": container with ID starting with 98b097381886815d101845570976bf1ee51d9ffee4aff5dea75ed775ba801be9 not found: ID does not exist" Sep 29 19:56:19 crc kubenswrapper[4741]: I0929 19:56:19.105831 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" path="/var/lib/kubelet/pods/6983b04b-ca22-4509-ad2d-c8d067a2709f/volumes" Sep 29 19:56:31 crc kubenswrapper[4741]: I0929 19:56:31.738445 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:56:31 crc kubenswrapper[4741]: I0929 19:56:31.738911 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:57:01 crc kubenswrapper[4741]: I0929 19:57:01.739326 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:57:01 crc kubenswrapper[4741]: I0929 19:57:01.740165 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 19:57:01 crc kubenswrapper[4741]: I0929 19:57:01.740248 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 19:57:01 crc kubenswrapper[4741]: I0929 19:57:01.741257 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89ba566aa7ea8f6245d4f453a81ad158f7f9415979df78f7544100d5299c5ac0"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 19:57:01 crc kubenswrapper[4741]: I0929 19:57:01.741362 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://89ba566aa7ea8f6245d4f453a81ad158f7f9415979df78f7544100d5299c5ac0" gracePeriod=600 Sep 29 19:57:02 crc kubenswrapper[4741]: I0929 19:57:02.623879 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="89ba566aa7ea8f6245d4f453a81ad158f7f9415979df78f7544100d5299c5ac0" exitCode=0 Sep 29 19:57:02 crc kubenswrapper[4741]: I0929 19:57:02.623959 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"89ba566aa7ea8f6245d4f453a81ad158f7f9415979df78f7544100d5299c5ac0"} Sep 29 19:57:02 crc kubenswrapper[4741]: I0929 19:57:02.624354 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb"} Sep 29 19:57:02 crc kubenswrapper[4741]: I0929 19:57:02.624379 4741 scope.go:117] "RemoveContainer" containerID="5732d43d68a93180b2a8de03b13840a22d26cc89d7312ce090c0aaf7b2af3cff" Sep 29 19:59:31 crc kubenswrapper[4741]: I0929 19:59:31.738421 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 19:59:31 crc kubenswrapper[4741]: I0929 19:59:31.739067 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.179282 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg"] Sep 29 20:00:00 crc kubenswrapper[4741]: E0929 20:00:00.180259 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="registry-server" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.180281 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="registry-server" Sep 29 20:00:00 crc kubenswrapper[4741]: E0929 20:00:00.180320 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="extract-utilities" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.180331 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="extract-utilities" Sep 29 20:00:00 crc kubenswrapper[4741]: E0929 20:00:00.180346 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="extract-content" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.180359 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="extract-content" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.180586 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6983b04b-ca22-4509-ad2d-c8d067a2709f" containerName="registry-server" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.181162 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.183852 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.184077 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.190022 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg"] Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.319493 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x88q\" (UniqueName: \"kubernetes.io/projected/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-kube-api-access-6x88q\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.319593 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-secret-volume\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.319627 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-config-volume\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.420637 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-secret-volume\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.420703 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-config-volume\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.420781 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x88q\" (UniqueName: \"kubernetes.io/projected/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-kube-api-access-6x88q\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.422136 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-config-volume\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.427286 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-secret-volume\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.441085 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x88q\" (UniqueName: \"kubernetes.io/projected/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-kube-api-access-6x88q\") pod \"collect-profiles-29319600-8fhqg\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.517339 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:00 crc kubenswrapper[4741]: I0929 20:00:00.947875 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg"] Sep 29 20:00:00 crc kubenswrapper[4741]: W0929 20:00:00.952752 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c9a64fc_57a2_49a6_bccc_25a23ae2835a.slice/crio-71297e4722f21103219d5b6fc4759542f3e0ff63080dfebe7d1d1e3a8cf3c269 WatchSource:0}: Error finding container 71297e4722f21103219d5b6fc4759542f3e0ff63080dfebe7d1d1e3a8cf3c269: Status 404 returned error can't find the container with id 71297e4722f21103219d5b6fc4759542f3e0ff63080dfebe7d1d1e3a8cf3c269 Sep 29 20:00:01 crc kubenswrapper[4741]: I0929 20:00:01.096731 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" event={"ID":"1c9a64fc-57a2-49a6-bccc-25a23ae2835a","Type":"ContainerStarted","Data":"71297e4722f21103219d5b6fc4759542f3e0ff63080dfebe7d1d1e3a8cf3c269"} Sep 29 20:00:01 crc kubenswrapper[4741]: I0929 20:00:01.739141 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:00:01 crc kubenswrapper[4741]: I0929 20:00:01.739651 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:00:02 crc kubenswrapper[4741]: I0929 20:00:02.108142 4741 generic.go:334] "Generic (PLEG): container finished" podID="1c9a64fc-57a2-49a6-bccc-25a23ae2835a" containerID="4d82aae8ccf2e5120dc9475a95d58491a15b05a856d163470bcbe7c53bb56db6" exitCode=0 Sep 29 20:00:02 crc kubenswrapper[4741]: I0929 20:00:02.108223 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" event={"ID":"1c9a64fc-57a2-49a6-bccc-25a23ae2835a","Type":"ContainerDied","Data":"4d82aae8ccf2e5120dc9475a95d58491a15b05a856d163470bcbe7c53bb56db6"} Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.376449 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.561201 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x88q\" (UniqueName: \"kubernetes.io/projected/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-kube-api-access-6x88q\") pod \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.561313 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-secret-volume\") pod \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.561669 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-config-volume\") pod \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\" (UID: \"1c9a64fc-57a2-49a6-bccc-25a23ae2835a\") " Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.562647 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-config-volume" (OuterVolumeSpecName: "config-volume") pod "1c9a64fc-57a2-49a6-bccc-25a23ae2835a" (UID: "1c9a64fc-57a2-49a6-bccc-25a23ae2835a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.568086 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-kube-api-access-6x88q" (OuterVolumeSpecName: "kube-api-access-6x88q") pod "1c9a64fc-57a2-49a6-bccc-25a23ae2835a" (UID: "1c9a64fc-57a2-49a6-bccc-25a23ae2835a"). InnerVolumeSpecName "kube-api-access-6x88q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.568528 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1c9a64fc-57a2-49a6-bccc-25a23ae2835a" (UID: "1c9a64fc-57a2-49a6-bccc-25a23ae2835a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.662806 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.662836 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:00:03 crc kubenswrapper[4741]: I0929 20:00:03.662850 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x88q\" (UniqueName: \"kubernetes.io/projected/1c9a64fc-57a2-49a6-bccc-25a23ae2835a-kube-api-access-6x88q\") on node \"crc\" DevicePath \"\"" Sep 29 20:00:04 crc kubenswrapper[4741]: I0929 20:00:04.121872 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" event={"ID":"1c9a64fc-57a2-49a6-bccc-25a23ae2835a","Type":"ContainerDied","Data":"71297e4722f21103219d5b6fc4759542f3e0ff63080dfebe7d1d1e3a8cf3c269"} Sep 29 20:00:04 crc kubenswrapper[4741]: I0929 20:00:04.121919 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71297e4722f21103219d5b6fc4759542f3e0ff63080dfebe7d1d1e3a8cf3c269" Sep 29 20:00:04 crc kubenswrapper[4741]: I0929 20:00:04.121961 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg" Sep 29 20:00:04 crc kubenswrapper[4741]: I0929 20:00:04.453352 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d"] Sep 29 20:00:04 crc kubenswrapper[4741]: I0929 20:00:04.457718 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319555-knl4d"] Sep 29 20:00:05 crc kubenswrapper[4741]: I0929 20:00:05.096233 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ff83e8-1d40-4ae2-9878-24ed37e22c3e" path="/var/lib/kubelet/pods/e0ff83e8-1d40-4ae2-9878-24ed37e22c3e/volumes" Sep 29 20:00:25 crc kubenswrapper[4741]: I0929 20:00:25.417500 4741 scope.go:117] "RemoveContainer" containerID="ac53733f11090f6bc9ed473e41f09c141c1505335e4106baba8391c443a7c356" Sep 29 20:00:31 crc kubenswrapper[4741]: I0929 20:00:31.738999 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:00:31 crc kubenswrapper[4741]: I0929 20:00:31.739566 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:00:31 crc kubenswrapper[4741]: I0929 20:00:31.739604 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:00:31 crc kubenswrapper[4741]: I0929 20:00:31.740274 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:00:31 crc kubenswrapper[4741]: I0929 20:00:31.740327 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" gracePeriod=600 Sep 29 20:00:31 crc kubenswrapper[4741]: E0929 20:00:31.863692 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:00:32 crc kubenswrapper[4741]: I0929 20:00:32.361457 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" exitCode=0 Sep 29 20:00:32 crc kubenswrapper[4741]: I0929 20:00:32.361820 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb"} Sep 29 20:00:32 crc kubenswrapper[4741]: I0929 20:00:32.361903 4741 scope.go:117] "RemoveContainer" containerID="89ba566aa7ea8f6245d4f453a81ad158f7f9415979df78f7544100d5299c5ac0" Sep 29 20:00:32 crc kubenswrapper[4741]: I0929 20:00:32.362680 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:00:32 crc kubenswrapper[4741]: E0929 20:00:32.362971 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:00:45 crc kubenswrapper[4741]: I0929 20:00:45.086071 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:00:45 crc kubenswrapper[4741]: E0929 20:00:45.086829 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:00:57 crc kubenswrapper[4741]: I0929 20:00:57.086370 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:00:57 crc kubenswrapper[4741]: E0929 20:00:57.087013 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:01:08 crc kubenswrapper[4741]: I0929 20:01:08.086036 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:01:08 crc kubenswrapper[4741]: E0929 20:01:08.086935 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:01:19 crc kubenswrapper[4741]: I0929 20:01:19.095818 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:01:19 crc kubenswrapper[4741]: E0929 20:01:19.096803 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:01:30 crc kubenswrapper[4741]: I0929 20:01:30.086053 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:01:30 crc kubenswrapper[4741]: E0929 20:01:30.086924 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:01:43 crc kubenswrapper[4741]: I0929 20:01:43.085758 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:01:43 crc kubenswrapper[4741]: E0929 20:01:43.086547 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.715883 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-trrtd"] Sep 29 20:01:46 crc kubenswrapper[4741]: E0929 20:01:46.718118 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9a64fc-57a2-49a6-bccc-25a23ae2835a" containerName="collect-profiles" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.718231 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9a64fc-57a2-49a6-bccc-25a23ae2835a" containerName="collect-profiles" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.719474 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9a64fc-57a2-49a6-bccc-25a23ae2835a" containerName="collect-profiles" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.721189 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.756219 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-trrtd"] Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.869125 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-catalog-content\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.869227 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-utilities\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.869258 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s5kj\" (UniqueName: \"kubernetes.io/projected/05b74776-4e37-432c-a1ca-edd7e3adcfe0-kube-api-access-7s5kj\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.970498 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-utilities\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.970549 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s5kj\" (UniqueName: \"kubernetes.io/projected/05b74776-4e37-432c-a1ca-edd7e3adcfe0-kube-api-access-7s5kj\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.970606 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-catalog-content\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.971230 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-utilities\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:46 crc kubenswrapper[4741]: I0929 20:01:46.971285 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-catalog-content\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:47 crc kubenswrapper[4741]: I0929 20:01:47.004765 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s5kj\" (UniqueName: \"kubernetes.io/projected/05b74776-4e37-432c-a1ca-edd7e3adcfe0-kube-api-access-7s5kj\") pod \"certified-operators-trrtd\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:47 crc kubenswrapper[4741]: I0929 20:01:47.056514 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:47 crc kubenswrapper[4741]: I0929 20:01:47.629016 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-trrtd"] Sep 29 20:01:47 crc kubenswrapper[4741]: I0929 20:01:47.957711 4741 generic.go:334] "Generic (PLEG): container finished" podID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerID="04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7" exitCode=0 Sep 29 20:01:47 crc kubenswrapper[4741]: I0929 20:01:47.957864 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trrtd" event={"ID":"05b74776-4e37-432c-a1ca-edd7e3adcfe0","Type":"ContainerDied","Data":"04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7"} Sep 29 20:01:47 crc kubenswrapper[4741]: I0929 20:01:47.958916 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trrtd" event={"ID":"05b74776-4e37-432c-a1ca-edd7e3adcfe0","Type":"ContainerStarted","Data":"f1fdf0164ccf8513b1f721f27514967b669b02096fad6c9c859173198cfa6d0b"} Sep 29 20:01:47 crc kubenswrapper[4741]: I0929 20:01:47.960069 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:01:48 crc kubenswrapper[4741]: I0929 20:01:48.969914 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trrtd" event={"ID":"05b74776-4e37-432c-a1ca-edd7e3adcfe0","Type":"ContainerStarted","Data":"729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73"} Sep 29 20:01:49 crc kubenswrapper[4741]: I0929 20:01:49.984067 4741 generic.go:334] "Generic (PLEG): container finished" podID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerID="729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73" exitCode=0 Sep 29 20:01:49 crc kubenswrapper[4741]: I0929 20:01:49.984118 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trrtd" event={"ID":"05b74776-4e37-432c-a1ca-edd7e3adcfe0","Type":"ContainerDied","Data":"729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73"} Sep 29 20:01:50 crc kubenswrapper[4741]: I0929 20:01:50.993280 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trrtd" event={"ID":"05b74776-4e37-432c-a1ca-edd7e3adcfe0","Type":"ContainerStarted","Data":"9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9"} Sep 29 20:01:51 crc kubenswrapper[4741]: I0929 20:01:51.015353 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-trrtd" podStartSLOduration=2.548678213 podStartE2EDuration="5.015337454s" podCreationTimestamp="2025-09-29 20:01:46 +0000 UTC" firstStartedPulling="2025-09-29 20:01:47.959808003 +0000 UTC m=+3149.607597335" lastFinishedPulling="2025-09-29 20:01:50.426467204 +0000 UTC m=+3152.074256576" observedRunningTime="2025-09-29 20:01:51.010578464 +0000 UTC m=+3152.658367806" watchObservedRunningTime="2025-09-29 20:01:51.015337454 +0000 UTC m=+3152.663126786" Sep 29 20:01:57 crc kubenswrapper[4741]: I0929 20:01:57.056647 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:57 crc kubenswrapper[4741]: I0929 20:01:57.057093 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:57 crc kubenswrapper[4741]: I0929 20:01:57.086500 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:01:57 crc kubenswrapper[4741]: E0929 20:01:57.086882 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:01:57 crc kubenswrapper[4741]: I0929 20:01:57.103336 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:58 crc kubenswrapper[4741]: I0929 20:01:58.089621 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:01:58 crc kubenswrapper[4741]: I0929 20:01:58.132360 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-trrtd"] Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.061716 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-trrtd" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="registry-server" containerID="cri-o://9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9" gracePeriod=2 Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.492783 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.573425 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s5kj\" (UniqueName: \"kubernetes.io/projected/05b74776-4e37-432c-a1ca-edd7e3adcfe0-kube-api-access-7s5kj\") pod \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.573506 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-catalog-content\") pod \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.573573 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-utilities\") pod \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\" (UID: \"05b74776-4e37-432c-a1ca-edd7e3adcfe0\") " Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.574561 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-utilities" (OuterVolumeSpecName: "utilities") pod "05b74776-4e37-432c-a1ca-edd7e3adcfe0" (UID: "05b74776-4e37-432c-a1ca-edd7e3adcfe0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.580250 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05b74776-4e37-432c-a1ca-edd7e3adcfe0-kube-api-access-7s5kj" (OuterVolumeSpecName: "kube-api-access-7s5kj") pod "05b74776-4e37-432c-a1ca-edd7e3adcfe0" (UID: "05b74776-4e37-432c-a1ca-edd7e3adcfe0"). InnerVolumeSpecName "kube-api-access-7s5kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.629508 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05b74776-4e37-432c-a1ca-edd7e3adcfe0" (UID: "05b74776-4e37-432c-a1ca-edd7e3adcfe0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.675031 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s5kj\" (UniqueName: \"kubernetes.io/projected/05b74776-4e37-432c-a1ca-edd7e3adcfe0-kube-api-access-7s5kj\") on node \"crc\" DevicePath \"\"" Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.675068 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:02:00 crc kubenswrapper[4741]: I0929 20:02:00.675081 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05b74776-4e37-432c-a1ca-edd7e3adcfe0-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.078402 4741 generic.go:334] "Generic (PLEG): container finished" podID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerID="9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9" exitCode=0 Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.078442 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trrtd" event={"ID":"05b74776-4e37-432c-a1ca-edd7e3adcfe0","Type":"ContainerDied","Data":"9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9"} Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.078467 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trrtd" event={"ID":"05b74776-4e37-432c-a1ca-edd7e3adcfe0","Type":"ContainerDied","Data":"f1fdf0164ccf8513b1f721f27514967b669b02096fad6c9c859173198cfa6d0b"} Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.078482 4741 scope.go:117] "RemoveContainer" containerID="9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.078529 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trrtd" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.111433 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-trrtd"] Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.115187 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-trrtd"] Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.121170 4741 scope.go:117] "RemoveContainer" containerID="729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.138106 4741 scope.go:117] "RemoveContainer" containerID="04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.163069 4741 scope.go:117] "RemoveContainer" containerID="9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9" Sep 29 20:02:01 crc kubenswrapper[4741]: E0929 20:02:01.163608 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9\": container with ID starting with 9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9 not found: ID does not exist" containerID="9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.163679 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9"} err="failed to get container status \"9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9\": rpc error: code = NotFound desc = could not find container \"9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9\": container with ID starting with 9e5338e0fac99bb5492f3e0e5dcb98c156377fc94ce3e254cbf3402d3cde10a9 not found: ID does not exist" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.163714 4741 scope.go:117] "RemoveContainer" containerID="729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73" Sep 29 20:02:01 crc kubenswrapper[4741]: E0929 20:02:01.164092 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73\": container with ID starting with 729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73 not found: ID does not exist" containerID="729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.164129 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73"} err="failed to get container status \"729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73\": rpc error: code = NotFound desc = could not find container \"729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73\": container with ID starting with 729caebc6f863b79e13c7df10efcf3b55719a06851a64a369e02fd8b8ce85a73 not found: ID does not exist" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.164154 4741 scope.go:117] "RemoveContainer" containerID="04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7" Sep 29 20:02:01 crc kubenswrapper[4741]: E0929 20:02:01.164548 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7\": container with ID starting with 04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7 not found: ID does not exist" containerID="04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7" Sep 29 20:02:01 crc kubenswrapper[4741]: I0929 20:02:01.164583 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7"} err="failed to get container status \"04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7\": rpc error: code = NotFound desc = could not find container \"04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7\": container with ID starting with 04ed3bfe129a04da88311316c309a27a57a2705eae6f316c0bad70d5b4b459c7 not found: ID does not exist" Sep 29 20:02:03 crc kubenswrapper[4741]: I0929 20:02:03.098842 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" path="/var/lib/kubelet/pods/05b74776-4e37-432c-a1ca-edd7e3adcfe0/volumes" Sep 29 20:02:12 crc kubenswrapper[4741]: I0929 20:02:12.086287 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:02:12 crc kubenswrapper[4741]: E0929 20:02:12.086906 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:02:24 crc kubenswrapper[4741]: I0929 20:02:24.085840 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:02:24 crc kubenswrapper[4741]: E0929 20:02:24.087118 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:02:39 crc kubenswrapper[4741]: I0929 20:02:39.090355 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:02:39 crc kubenswrapper[4741]: E0929 20:02:39.091478 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:02:54 crc kubenswrapper[4741]: I0929 20:02:54.085843 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:02:54 crc kubenswrapper[4741]: E0929 20:02:54.087792 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:03:08 crc kubenswrapper[4741]: I0929 20:03:08.085426 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:03:08 crc kubenswrapper[4741]: E0929 20:03:08.086140 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:03:19 crc kubenswrapper[4741]: I0929 20:03:19.091559 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:03:19 crc kubenswrapper[4741]: E0929 20:03:19.092441 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:03:32 crc kubenswrapper[4741]: I0929 20:03:32.086527 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:03:32 crc kubenswrapper[4741]: E0929 20:03:32.087291 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:03:46 crc kubenswrapper[4741]: I0929 20:03:46.086713 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:03:46 crc kubenswrapper[4741]: E0929 20:03:46.087725 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:03:57 crc kubenswrapper[4741]: I0929 20:03:57.086560 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:03:57 crc kubenswrapper[4741]: E0929 20:03:57.087632 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:04:11 crc kubenswrapper[4741]: I0929 20:04:11.088265 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:04:11 crc kubenswrapper[4741]: E0929 20:04:11.090211 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.939829 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sv2h7"] Sep 29 20:04:19 crc kubenswrapper[4741]: E0929 20:04:19.940576 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="extract-content" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.940588 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="extract-content" Sep 29 20:04:19 crc kubenswrapper[4741]: E0929 20:04:19.940601 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="registry-server" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.940608 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="registry-server" Sep 29 20:04:19 crc kubenswrapper[4741]: E0929 20:04:19.940640 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="extract-utilities" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.940646 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="extract-utilities" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.940767 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="05b74776-4e37-432c-a1ca-edd7e3adcfe0" containerName="registry-server" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.941873 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.945748 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sv2h7"] Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.979498 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sjsv\" (UniqueName: \"kubernetes.io/projected/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-kube-api-access-8sjsv\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.979962 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-catalog-content\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:19 crc kubenswrapper[4741]: I0929 20:04:19.980000 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-utilities\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.080552 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-catalog-content\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.080605 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-utilities\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.080630 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sjsv\" (UniqueName: \"kubernetes.io/projected/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-kube-api-access-8sjsv\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.081065 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-catalog-content\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.081179 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-utilities\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.101710 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sjsv\" (UniqueName: \"kubernetes.io/projected/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-kube-api-access-8sjsv\") pod \"redhat-marketplace-sv2h7\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.259041 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:20 crc kubenswrapper[4741]: I0929 20:04:20.540601 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sv2h7"] Sep 29 20:04:21 crc kubenswrapper[4741]: I0929 20:04:21.166134 4741 generic.go:334] "Generic (PLEG): container finished" podID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerID="2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd" exitCode=0 Sep 29 20:04:21 crc kubenswrapper[4741]: I0929 20:04:21.166232 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sv2h7" event={"ID":"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8","Type":"ContainerDied","Data":"2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd"} Sep 29 20:04:21 crc kubenswrapper[4741]: I0929 20:04:21.166308 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sv2h7" event={"ID":"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8","Type":"ContainerStarted","Data":"432e6b53f5c44b933f5d306dbf607612f82a6185a6bee76849ffee31e647a773"} Sep 29 20:04:23 crc kubenswrapper[4741]: I0929 20:04:23.185618 4741 generic.go:334] "Generic (PLEG): container finished" podID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerID="aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179" exitCode=0 Sep 29 20:04:23 crc kubenswrapper[4741]: I0929 20:04:23.185731 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sv2h7" event={"ID":"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8","Type":"ContainerDied","Data":"aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179"} Sep 29 20:04:24 crc kubenswrapper[4741]: I0929 20:04:24.085498 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:04:24 crc kubenswrapper[4741]: E0929 20:04:24.085928 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:04:24 crc kubenswrapper[4741]: I0929 20:04:24.196822 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sv2h7" event={"ID":"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8","Type":"ContainerStarted","Data":"ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f"} Sep 29 20:04:24 crc kubenswrapper[4741]: I0929 20:04:24.214854 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sv2h7" podStartSLOduration=2.662208322 podStartE2EDuration="5.214833009s" podCreationTimestamp="2025-09-29 20:04:19 +0000 UTC" firstStartedPulling="2025-09-29 20:04:21.167650113 +0000 UTC m=+3302.815439445" lastFinishedPulling="2025-09-29 20:04:23.7202748 +0000 UTC m=+3305.368064132" observedRunningTime="2025-09-29 20:04:24.213091744 +0000 UTC m=+3305.860881086" watchObservedRunningTime="2025-09-29 20:04:24.214833009 +0000 UTC m=+3305.862622341" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.335024 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-557xn"] Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.337015 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.346139 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-557xn"] Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.363892 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n74gv\" (UniqueName: \"kubernetes.io/projected/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-kube-api-access-n74gv\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.363977 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-catalog-content\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.364031 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-utilities\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.465060 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n74gv\" (UniqueName: \"kubernetes.io/projected/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-kube-api-access-n74gv\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.465127 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-catalog-content\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.465166 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-utilities\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.465797 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-utilities\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.465803 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-catalog-content\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.485750 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n74gv\" (UniqueName: \"kubernetes.io/projected/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-kube-api-access-n74gv\") pod \"redhat-operators-557xn\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:25 crc kubenswrapper[4741]: I0929 20:04:25.672946 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:26 crc kubenswrapper[4741]: I0929 20:04:26.104562 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-557xn"] Sep 29 20:04:26 crc kubenswrapper[4741]: I0929 20:04:26.212340 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-557xn" event={"ID":"43e02aa2-bd36-42bb-ae6a-bded2e4729ae","Type":"ContainerStarted","Data":"41503d75ec8cbf4568f34f164a99a60feab44f51ed4b1395e39b5603ad65c9c9"} Sep 29 20:04:27 crc kubenswrapper[4741]: I0929 20:04:27.220918 4741 generic.go:334] "Generic (PLEG): container finished" podID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerID="85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964" exitCode=0 Sep 29 20:04:27 crc kubenswrapper[4741]: I0929 20:04:27.220975 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-557xn" event={"ID":"43e02aa2-bd36-42bb-ae6a-bded2e4729ae","Type":"ContainerDied","Data":"85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964"} Sep 29 20:04:29 crc kubenswrapper[4741]: I0929 20:04:29.241445 4741 generic.go:334] "Generic (PLEG): container finished" podID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerID="09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef" exitCode=0 Sep 29 20:04:29 crc kubenswrapper[4741]: I0929 20:04:29.241496 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-557xn" event={"ID":"43e02aa2-bd36-42bb-ae6a-bded2e4729ae","Type":"ContainerDied","Data":"09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef"} Sep 29 20:04:30 crc kubenswrapper[4741]: I0929 20:04:30.249472 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-557xn" event={"ID":"43e02aa2-bd36-42bb-ae6a-bded2e4729ae","Type":"ContainerStarted","Data":"b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404"} Sep 29 20:04:30 crc kubenswrapper[4741]: I0929 20:04:30.259465 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:30 crc kubenswrapper[4741]: I0929 20:04:30.259820 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:30 crc kubenswrapper[4741]: I0929 20:04:30.305507 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:30 crc kubenswrapper[4741]: I0929 20:04:30.332103 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-557xn" podStartSLOduration=2.62317216 podStartE2EDuration="5.332087146s" podCreationTimestamp="2025-09-29 20:04:25 +0000 UTC" firstStartedPulling="2025-09-29 20:04:27.223325414 +0000 UTC m=+3308.871114776" lastFinishedPulling="2025-09-29 20:04:29.93224039 +0000 UTC m=+3311.580029762" observedRunningTime="2025-09-29 20:04:30.274589079 +0000 UTC m=+3311.922378481" watchObservedRunningTime="2025-09-29 20:04:30.332087146 +0000 UTC m=+3311.979876478" Sep 29 20:04:31 crc kubenswrapper[4741]: I0929 20:04:31.308344 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:32 crc kubenswrapper[4741]: I0929 20:04:32.508688 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sv2h7"] Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.272534 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sv2h7" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="registry-server" containerID="cri-o://ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f" gracePeriod=2 Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.758833 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.807829 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-utilities\") pod \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.807882 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-catalog-content\") pod \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.807985 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sjsv\" (UniqueName: \"kubernetes.io/projected/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-kube-api-access-8sjsv\") pod \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\" (UID: \"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8\") " Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.808620 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-utilities" (OuterVolumeSpecName: "utilities") pod "91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" (UID: "91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.813730 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-kube-api-access-8sjsv" (OuterVolumeSpecName: "kube-api-access-8sjsv") pod "91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" (UID: "91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8"). InnerVolumeSpecName "kube-api-access-8sjsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.836536 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" (UID: "91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.909158 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.909201 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:04:33 crc kubenswrapper[4741]: I0929 20:04:33.909215 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sjsv\" (UniqueName: \"kubernetes.io/projected/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8-kube-api-access-8sjsv\") on node \"crc\" DevicePath \"\"" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.281059 4741 generic.go:334] "Generic (PLEG): container finished" podID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerID="ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f" exitCode=0 Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.281194 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sv2h7" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.281215 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sv2h7" event={"ID":"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8","Type":"ContainerDied","Data":"ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f"} Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.281624 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sv2h7" event={"ID":"91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8","Type":"ContainerDied","Data":"432e6b53f5c44b933f5d306dbf607612f82a6185a6bee76849ffee31e647a773"} Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.281649 4741 scope.go:117] "RemoveContainer" containerID="ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.300949 4741 scope.go:117] "RemoveContainer" containerID="aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.316931 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sv2h7"] Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.321064 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sv2h7"] Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.338017 4741 scope.go:117] "RemoveContainer" containerID="2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.352866 4741 scope.go:117] "RemoveContainer" containerID="ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f" Sep 29 20:04:34 crc kubenswrapper[4741]: E0929 20:04:34.353432 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f\": container with ID starting with ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f not found: ID does not exist" containerID="ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.353484 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f"} err="failed to get container status \"ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f\": rpc error: code = NotFound desc = could not find container \"ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f\": container with ID starting with ace0ff7a3c91166baa62bde725033e6391097a6bbd5ad12b1039cbf13021777f not found: ID does not exist" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.353557 4741 scope.go:117] "RemoveContainer" containerID="aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179" Sep 29 20:04:34 crc kubenswrapper[4741]: E0929 20:04:34.353954 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179\": container with ID starting with aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179 not found: ID does not exist" containerID="aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.354063 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179"} err="failed to get container status \"aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179\": rpc error: code = NotFound desc = could not find container \"aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179\": container with ID starting with aa48a8a028ad50a5618395a70ef9052aad398bd8985fdd21f1e4057421e9d179 not found: ID does not exist" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.354143 4741 scope.go:117] "RemoveContainer" containerID="2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd" Sep 29 20:04:34 crc kubenswrapper[4741]: E0929 20:04:34.354448 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd\": container with ID starting with 2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd not found: ID does not exist" containerID="2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd" Sep 29 20:04:34 crc kubenswrapper[4741]: I0929 20:04:34.354541 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd"} err="failed to get container status \"2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd\": rpc error: code = NotFound desc = could not find container \"2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd\": container with ID starting with 2a58387fa346246174353fd5b756939371463e2ccecd36e7869624f87d2877dd not found: ID does not exist" Sep 29 20:04:35 crc kubenswrapper[4741]: I0929 20:04:35.086643 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:04:35 crc kubenswrapper[4741]: E0929 20:04:35.087049 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:04:35 crc kubenswrapper[4741]: I0929 20:04:35.102065 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" path="/var/lib/kubelet/pods/91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8/volumes" Sep 29 20:04:35 crc kubenswrapper[4741]: I0929 20:04:35.673561 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:35 crc kubenswrapper[4741]: I0929 20:04:35.673613 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:35 crc kubenswrapper[4741]: I0929 20:04:35.718024 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:36 crc kubenswrapper[4741]: I0929 20:04:36.359983 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:36 crc kubenswrapper[4741]: I0929 20:04:36.918053 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-557xn"] Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.323022 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-557xn" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="registry-server" containerID="cri-o://b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404" gracePeriod=2 Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.691340 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.779787 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-catalog-content\") pod \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.779892 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-utilities\") pod \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.779997 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n74gv\" (UniqueName: \"kubernetes.io/projected/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-kube-api-access-n74gv\") pod \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\" (UID: \"43e02aa2-bd36-42bb-ae6a-bded2e4729ae\") " Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.781059 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-utilities" (OuterVolumeSpecName: "utilities") pod "43e02aa2-bd36-42bb-ae6a-bded2e4729ae" (UID: "43e02aa2-bd36-42bb-ae6a-bded2e4729ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.784860 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-kube-api-access-n74gv" (OuterVolumeSpecName: "kube-api-access-n74gv") pod "43e02aa2-bd36-42bb-ae6a-bded2e4729ae" (UID: "43e02aa2-bd36-42bb-ae6a-bded2e4729ae"). InnerVolumeSpecName "kube-api-access-n74gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.881594 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:04:38 crc kubenswrapper[4741]: I0929 20:04:38.881638 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n74gv\" (UniqueName: \"kubernetes.io/projected/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-kube-api-access-n74gv\") on node \"crc\" DevicePath \"\"" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.332983 4741 generic.go:334] "Generic (PLEG): container finished" podID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerID="b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404" exitCode=0 Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.333031 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-557xn" event={"ID":"43e02aa2-bd36-42bb-ae6a-bded2e4729ae","Type":"ContainerDied","Data":"b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404"} Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.333066 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-557xn" event={"ID":"43e02aa2-bd36-42bb-ae6a-bded2e4729ae","Type":"ContainerDied","Data":"41503d75ec8cbf4568f34f164a99a60feab44f51ed4b1395e39b5603ad65c9c9"} Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.333083 4741 scope.go:117] "RemoveContainer" containerID="b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.333038 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-557xn" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.351486 4741 scope.go:117] "RemoveContainer" containerID="09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.373000 4741 scope.go:117] "RemoveContainer" containerID="85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.395265 4741 scope.go:117] "RemoveContainer" containerID="b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404" Sep 29 20:04:39 crc kubenswrapper[4741]: E0929 20:04:39.395686 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404\": container with ID starting with b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404 not found: ID does not exist" containerID="b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.395721 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404"} err="failed to get container status \"b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404\": rpc error: code = NotFound desc = could not find container \"b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404\": container with ID starting with b6004b38c5c75e7e8f6db75f58cd70db19c7e516366bc50136ef31d114a10404 not found: ID does not exist" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.395743 4741 scope.go:117] "RemoveContainer" containerID="09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef" Sep 29 20:04:39 crc kubenswrapper[4741]: E0929 20:04:39.396153 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef\": container with ID starting with 09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef not found: ID does not exist" containerID="09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.396179 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef"} err="failed to get container status \"09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef\": rpc error: code = NotFound desc = could not find container \"09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef\": container with ID starting with 09769bba987d44893c0f67e6cdfec62ccdfe6ee033f89e0478708fddf0c9bfef not found: ID does not exist" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.396195 4741 scope.go:117] "RemoveContainer" containerID="85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964" Sep 29 20:04:39 crc kubenswrapper[4741]: E0929 20:04:39.396448 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964\": container with ID starting with 85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964 not found: ID does not exist" containerID="85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.396467 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964"} err="failed to get container status \"85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964\": rpc error: code = NotFound desc = could not find container \"85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964\": container with ID starting with 85a2fd95a273d3c1c8f251bef1e918f44dd570bdcc63d5d7374936a3f4394964 not found: ID does not exist" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.789168 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43e02aa2-bd36-42bb-ae6a-bded2e4729ae" (UID: "43e02aa2-bd36-42bb-ae6a-bded2e4729ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.791962 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e02aa2-bd36-42bb-ae6a-bded2e4729ae-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.959754 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-557xn"] Sep 29 20:04:39 crc kubenswrapper[4741]: I0929 20:04:39.968631 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-557xn"] Sep 29 20:04:41 crc kubenswrapper[4741]: I0929 20:04:41.094155 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" path="/var/lib/kubelet/pods/43e02aa2-bd36-42bb-ae6a-bded2e4729ae/volumes" Sep 29 20:04:46 crc kubenswrapper[4741]: I0929 20:04:46.085773 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:04:46 crc kubenswrapper[4741]: E0929 20:04:46.086568 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:04:59 crc kubenswrapper[4741]: I0929 20:04:59.089111 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:04:59 crc kubenswrapper[4741]: E0929 20:04:59.089773 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:05:10 crc kubenswrapper[4741]: I0929 20:05:10.086113 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:05:10 crc kubenswrapper[4741]: E0929 20:05:10.087085 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:05:21 crc kubenswrapper[4741]: I0929 20:05:21.086655 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:05:21 crc kubenswrapper[4741]: E0929 20:05:21.087317 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:05:34 crc kubenswrapper[4741]: I0929 20:05:34.087000 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:05:34 crc kubenswrapper[4741]: I0929 20:05:34.746147 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"35e92b7732e29719fe07a9a393f0226f2393f56ecd021e5a4cd27179d979d479"} Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.235132 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-smsqw"] Sep 29 20:06:58 crc kubenswrapper[4741]: E0929 20:06:58.235994 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="extract-utilities" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236008 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="extract-utilities" Sep 29 20:06:58 crc kubenswrapper[4741]: E0929 20:06:58.236034 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="registry-server" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236042 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="registry-server" Sep 29 20:06:58 crc kubenswrapper[4741]: E0929 20:06:58.236061 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="extract-utilities" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236067 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="extract-utilities" Sep 29 20:06:58 crc kubenswrapper[4741]: E0929 20:06:58.236087 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="extract-content" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236093 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="extract-content" Sep 29 20:06:58 crc kubenswrapper[4741]: E0929 20:06:58.236112 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="registry-server" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236118 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="registry-server" Sep 29 20:06:58 crc kubenswrapper[4741]: E0929 20:06:58.236126 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="extract-content" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236132 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="extract-content" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236268 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e02aa2-bd36-42bb-ae6a-bded2e4729ae" containerName="registry-server" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.236294 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f8e416-9c8b-4f3a-8e9a-2a82aa31a4e8" containerName="registry-server" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.237299 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.249136 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smsqw"] Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.351254 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-utilities\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.351316 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-catalog-content\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.351374 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xxgl\" (UniqueName: \"kubernetes.io/projected/8128b3c2-1b10-4463-8159-19910d3168c7-kube-api-access-9xxgl\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.452558 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-utilities\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.452600 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-catalog-content\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.452629 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xxgl\" (UniqueName: \"kubernetes.io/projected/8128b3c2-1b10-4463-8159-19910d3168c7-kube-api-access-9xxgl\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.453208 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-catalog-content\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.453232 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-utilities\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.469664 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xxgl\" (UniqueName: \"kubernetes.io/projected/8128b3c2-1b10-4463-8159-19910d3168c7-kube-api-access-9xxgl\") pod \"community-operators-smsqw\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:58 crc kubenswrapper[4741]: I0929 20:06:58.565627 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:06:59 crc kubenswrapper[4741]: I0929 20:06:59.039011 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smsqw"] Sep 29 20:06:59 crc kubenswrapper[4741]: I0929 20:06:59.401004 4741 generic.go:334] "Generic (PLEG): container finished" podID="8128b3c2-1b10-4463-8159-19910d3168c7" containerID="1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3" exitCode=0 Sep 29 20:06:59 crc kubenswrapper[4741]: I0929 20:06:59.401106 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smsqw" event={"ID":"8128b3c2-1b10-4463-8159-19910d3168c7","Type":"ContainerDied","Data":"1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3"} Sep 29 20:06:59 crc kubenswrapper[4741]: I0929 20:06:59.401558 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smsqw" event={"ID":"8128b3c2-1b10-4463-8159-19910d3168c7","Type":"ContainerStarted","Data":"fa9dd84902c8f1032108b4c28ad037ea8784750c54195fb89d38e09776962539"} Sep 29 20:06:59 crc kubenswrapper[4741]: I0929 20:06:59.403518 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:07:01 crc kubenswrapper[4741]: I0929 20:07:01.416114 4741 generic.go:334] "Generic (PLEG): container finished" podID="8128b3c2-1b10-4463-8159-19910d3168c7" containerID="531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486" exitCode=0 Sep 29 20:07:01 crc kubenswrapper[4741]: I0929 20:07:01.416203 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smsqw" event={"ID":"8128b3c2-1b10-4463-8159-19910d3168c7","Type":"ContainerDied","Data":"531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486"} Sep 29 20:07:02 crc kubenswrapper[4741]: I0929 20:07:02.424781 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smsqw" event={"ID":"8128b3c2-1b10-4463-8159-19910d3168c7","Type":"ContainerStarted","Data":"3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff"} Sep 29 20:07:02 crc kubenswrapper[4741]: I0929 20:07:02.446138 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-smsqw" podStartSLOduration=1.6603991219999998 podStartE2EDuration="4.446121586s" podCreationTimestamp="2025-09-29 20:06:58 +0000 UTC" firstStartedPulling="2025-09-29 20:06:59.403193893 +0000 UTC m=+3461.050983225" lastFinishedPulling="2025-09-29 20:07:02.188916357 +0000 UTC m=+3463.836705689" observedRunningTime="2025-09-29 20:07:02.440372724 +0000 UTC m=+3464.088162056" watchObservedRunningTime="2025-09-29 20:07:02.446121586 +0000 UTC m=+3464.093910918" Sep 29 20:07:08 crc kubenswrapper[4741]: I0929 20:07:08.566701 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:07:08 crc kubenswrapper[4741]: I0929 20:07:08.566989 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:07:08 crc kubenswrapper[4741]: I0929 20:07:08.609189 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:07:09 crc kubenswrapper[4741]: I0929 20:07:09.565208 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:07:09 crc kubenswrapper[4741]: I0929 20:07:09.630234 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-smsqw"] Sep 29 20:07:11 crc kubenswrapper[4741]: I0929 20:07:11.502263 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-smsqw" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="registry-server" containerID="cri-o://3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff" gracePeriod=2 Sep 29 20:07:11 crc kubenswrapper[4741]: I0929 20:07:11.937379 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.043603 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-utilities\") pod \"8128b3c2-1b10-4463-8159-19910d3168c7\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.043716 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-catalog-content\") pod \"8128b3c2-1b10-4463-8159-19910d3168c7\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.043942 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xxgl\" (UniqueName: \"kubernetes.io/projected/8128b3c2-1b10-4463-8159-19910d3168c7-kube-api-access-9xxgl\") pod \"8128b3c2-1b10-4463-8159-19910d3168c7\" (UID: \"8128b3c2-1b10-4463-8159-19910d3168c7\") " Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.045951 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-utilities" (OuterVolumeSpecName: "utilities") pod "8128b3c2-1b10-4463-8159-19910d3168c7" (UID: "8128b3c2-1b10-4463-8159-19910d3168c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.051485 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8128b3c2-1b10-4463-8159-19910d3168c7-kube-api-access-9xxgl" (OuterVolumeSpecName: "kube-api-access-9xxgl") pod "8128b3c2-1b10-4463-8159-19910d3168c7" (UID: "8128b3c2-1b10-4463-8159-19910d3168c7"). InnerVolumeSpecName "kube-api-access-9xxgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.092740 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8128b3c2-1b10-4463-8159-19910d3168c7" (UID: "8128b3c2-1b10-4463-8159-19910d3168c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.145576 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.145615 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8128b3c2-1b10-4463-8159-19910d3168c7-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.145630 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xxgl\" (UniqueName: \"kubernetes.io/projected/8128b3c2-1b10-4463-8159-19910d3168c7-kube-api-access-9xxgl\") on node \"crc\" DevicePath \"\"" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.511846 4741 generic.go:334] "Generic (PLEG): container finished" podID="8128b3c2-1b10-4463-8159-19910d3168c7" containerID="3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff" exitCode=0 Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.511911 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smsqw" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.511907 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smsqw" event={"ID":"8128b3c2-1b10-4463-8159-19910d3168c7","Type":"ContainerDied","Data":"3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff"} Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.511990 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smsqw" event={"ID":"8128b3c2-1b10-4463-8159-19910d3168c7","Type":"ContainerDied","Data":"fa9dd84902c8f1032108b4c28ad037ea8784750c54195fb89d38e09776962539"} Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.512024 4741 scope.go:117] "RemoveContainer" containerID="3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.533140 4741 scope.go:117] "RemoveContainer" containerID="531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.551521 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-smsqw"] Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.557868 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-smsqw"] Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.570638 4741 scope.go:117] "RemoveContainer" containerID="1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.589067 4741 scope.go:117] "RemoveContainer" containerID="3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff" Sep 29 20:07:12 crc kubenswrapper[4741]: E0929 20:07:12.589534 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff\": container with ID starting with 3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff not found: ID does not exist" containerID="3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.589576 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff"} err="failed to get container status \"3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff\": rpc error: code = NotFound desc = could not find container \"3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff\": container with ID starting with 3a55af8b0d843712703841bbd851035f170697ae5fbd0f789cebd3d412a0f7ff not found: ID does not exist" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.589605 4741 scope.go:117] "RemoveContainer" containerID="531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486" Sep 29 20:07:12 crc kubenswrapper[4741]: E0929 20:07:12.589994 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486\": container with ID starting with 531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486 not found: ID does not exist" containerID="531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.590043 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486"} err="failed to get container status \"531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486\": rpc error: code = NotFound desc = could not find container \"531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486\": container with ID starting with 531f82eb3736f357d305fe0ee40c626968813348c007405b90215382888d7486 not found: ID does not exist" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.590075 4741 scope.go:117] "RemoveContainer" containerID="1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3" Sep 29 20:07:12 crc kubenswrapper[4741]: E0929 20:07:12.590339 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3\": container with ID starting with 1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3 not found: ID does not exist" containerID="1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3" Sep 29 20:07:12 crc kubenswrapper[4741]: I0929 20:07:12.590360 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3"} err="failed to get container status \"1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3\": rpc error: code = NotFound desc = could not find container \"1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3\": container with ID starting with 1a123f4dc7f2a757ff0b77f41354a48337d485b923c0db321a8f9e2d7ddcc5d3 not found: ID does not exist" Sep 29 20:07:13 crc kubenswrapper[4741]: I0929 20:07:13.111627 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" path="/var/lib/kubelet/pods/8128b3c2-1b10-4463-8159-19910d3168c7/volumes" Sep 29 20:08:01 crc kubenswrapper[4741]: I0929 20:08:01.738803 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:08:01 crc kubenswrapper[4741]: I0929 20:08:01.739479 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:08:31 crc kubenswrapper[4741]: I0929 20:08:31.738895 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:08:31 crc kubenswrapper[4741]: I0929 20:08:31.739449 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:09:01 crc kubenswrapper[4741]: I0929 20:09:01.738461 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:09:01 crc kubenswrapper[4741]: I0929 20:09:01.739082 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:09:01 crc kubenswrapper[4741]: I0929 20:09:01.739135 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:09:01 crc kubenswrapper[4741]: I0929 20:09:01.739912 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"35e92b7732e29719fe07a9a393f0226f2393f56ecd021e5a4cd27179d979d479"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:09:01 crc kubenswrapper[4741]: I0929 20:09:01.739978 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://35e92b7732e29719fe07a9a393f0226f2393f56ecd021e5a4cd27179d979d479" gracePeriod=600 Sep 29 20:09:02 crc kubenswrapper[4741]: I0929 20:09:02.297343 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="35e92b7732e29719fe07a9a393f0226f2393f56ecd021e5a4cd27179d979d479" exitCode=0 Sep 29 20:09:02 crc kubenswrapper[4741]: I0929 20:09:02.297447 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"35e92b7732e29719fe07a9a393f0226f2393f56ecd021e5a4cd27179d979d479"} Sep 29 20:09:02 crc kubenswrapper[4741]: I0929 20:09:02.297699 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8"} Sep 29 20:09:02 crc kubenswrapper[4741]: I0929 20:09:02.297726 4741 scope.go:117] "RemoveContainer" containerID="4134dea8169c31082b7b01609b259744ee9d9e1cebdc56f952d9cda5dfc97eeb" Sep 29 20:11:31 crc kubenswrapper[4741]: I0929 20:11:31.738577 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:11:31 crc kubenswrapper[4741]: I0929 20:11:31.739237 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.291171 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5j4kk"] Sep 29 20:11:51 crc kubenswrapper[4741]: E0929 20:11:51.292548 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="registry-server" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.292583 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="registry-server" Sep 29 20:11:51 crc kubenswrapper[4741]: E0929 20:11:51.292614 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="extract-content" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.292656 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="extract-content" Sep 29 20:11:51 crc kubenswrapper[4741]: E0929 20:11:51.292686 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="extract-utilities" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.292703 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="extract-utilities" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.293061 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8128b3c2-1b10-4463-8159-19910d3168c7" containerName="registry-server" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.296936 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.306782 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j4kk"] Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.421819 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-utilities\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.422094 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-catalog-content\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.422166 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l98dg\" (UniqueName: \"kubernetes.io/projected/04020ceb-cea7-484f-bd11-ff4546937ced-kube-api-access-l98dg\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.523480 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-catalog-content\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.523554 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l98dg\" (UniqueName: \"kubernetes.io/projected/04020ceb-cea7-484f-bd11-ff4546937ced-kube-api-access-l98dg\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.523620 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-utilities\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.524221 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-utilities\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.524488 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-catalog-content\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.557230 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l98dg\" (UniqueName: \"kubernetes.io/projected/04020ceb-cea7-484f-bd11-ff4546937ced-kube-api-access-l98dg\") pod \"certified-operators-5j4kk\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:51 crc kubenswrapper[4741]: I0929 20:11:51.625008 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:11:52 crc kubenswrapper[4741]: I0929 20:11:52.096081 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j4kk"] Sep 29 20:11:52 crc kubenswrapper[4741]: I0929 20:11:52.618793 4741 generic.go:334] "Generic (PLEG): container finished" podID="04020ceb-cea7-484f-bd11-ff4546937ced" containerID="54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a" exitCode=0 Sep 29 20:11:52 crc kubenswrapper[4741]: I0929 20:11:52.618841 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j4kk" event={"ID":"04020ceb-cea7-484f-bd11-ff4546937ced","Type":"ContainerDied","Data":"54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a"} Sep 29 20:11:52 crc kubenswrapper[4741]: I0929 20:11:52.619006 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j4kk" event={"ID":"04020ceb-cea7-484f-bd11-ff4546937ced","Type":"ContainerStarted","Data":"d7801457ab8f92d4b2ee2a696e9c041837d183ebd0a24c2704266e85bd3734a3"} Sep 29 20:11:54 crc kubenswrapper[4741]: I0929 20:11:54.639512 4741 generic.go:334] "Generic (PLEG): container finished" podID="04020ceb-cea7-484f-bd11-ff4546937ced" containerID="0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45" exitCode=0 Sep 29 20:11:54 crc kubenswrapper[4741]: I0929 20:11:54.639691 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j4kk" event={"ID":"04020ceb-cea7-484f-bd11-ff4546937ced","Type":"ContainerDied","Data":"0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45"} Sep 29 20:11:55 crc kubenswrapper[4741]: I0929 20:11:55.655324 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j4kk" event={"ID":"04020ceb-cea7-484f-bd11-ff4546937ced","Type":"ContainerStarted","Data":"ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b"} Sep 29 20:11:55 crc kubenswrapper[4741]: I0929 20:11:55.684495 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5j4kk" podStartSLOduration=2.177969058 podStartE2EDuration="4.68447145s" podCreationTimestamp="2025-09-29 20:11:51 +0000 UTC" firstStartedPulling="2025-09-29 20:11:52.621859504 +0000 UTC m=+3754.269648846" lastFinishedPulling="2025-09-29 20:11:55.128361896 +0000 UTC m=+3756.776151238" observedRunningTime="2025-09-29 20:11:55.679078846 +0000 UTC m=+3757.326868188" watchObservedRunningTime="2025-09-29 20:11:55.68447145 +0000 UTC m=+3757.332260792" Sep 29 20:12:01 crc kubenswrapper[4741]: I0929 20:12:01.625867 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:12:01 crc kubenswrapper[4741]: I0929 20:12:01.626232 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:12:01 crc kubenswrapper[4741]: I0929 20:12:01.675240 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:12:01 crc kubenswrapper[4741]: I0929 20:12:01.739023 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:12:01 crc kubenswrapper[4741]: I0929 20:12:01.739087 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:12:01 crc kubenswrapper[4741]: I0929 20:12:01.757226 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:12:01 crc kubenswrapper[4741]: I0929 20:12:01.917151 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5j4kk"] Sep 29 20:12:03 crc kubenswrapper[4741]: I0929 20:12:03.730537 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5j4kk" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="registry-server" containerID="cri-o://ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b" gracePeriod=2 Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.212220 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.314552 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-utilities\") pod \"04020ceb-cea7-484f-bd11-ff4546937ced\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.314738 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l98dg\" (UniqueName: \"kubernetes.io/projected/04020ceb-cea7-484f-bd11-ff4546937ced-kube-api-access-l98dg\") pod \"04020ceb-cea7-484f-bd11-ff4546937ced\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.314790 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-catalog-content\") pod \"04020ceb-cea7-484f-bd11-ff4546937ced\" (UID: \"04020ceb-cea7-484f-bd11-ff4546937ced\") " Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.320853 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04020ceb-cea7-484f-bd11-ff4546937ced-kube-api-access-l98dg" (OuterVolumeSpecName: "kube-api-access-l98dg") pod "04020ceb-cea7-484f-bd11-ff4546937ced" (UID: "04020ceb-cea7-484f-bd11-ff4546937ced"). InnerVolumeSpecName "kube-api-access-l98dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.321058 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-utilities" (OuterVolumeSpecName: "utilities") pod "04020ceb-cea7-484f-bd11-ff4546937ced" (UID: "04020ceb-cea7-484f-bd11-ff4546937ced"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.373495 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04020ceb-cea7-484f-bd11-ff4546937ced" (UID: "04020ceb-cea7-484f-bd11-ff4546937ced"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.416851 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l98dg\" (UniqueName: \"kubernetes.io/projected/04020ceb-cea7-484f-bd11-ff4546937ced-kube-api-access-l98dg\") on node \"crc\" DevicePath \"\"" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.416898 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.416913 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04020ceb-cea7-484f-bd11-ff4546937ced-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.744683 4741 generic.go:334] "Generic (PLEG): container finished" podID="04020ceb-cea7-484f-bd11-ff4546937ced" containerID="ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b" exitCode=0 Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.744753 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j4kk" event={"ID":"04020ceb-cea7-484f-bd11-ff4546937ced","Type":"ContainerDied","Data":"ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b"} Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.744780 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j4kk" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.744799 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j4kk" event={"ID":"04020ceb-cea7-484f-bd11-ff4546937ced","Type":"ContainerDied","Data":"d7801457ab8f92d4b2ee2a696e9c041837d183ebd0a24c2704266e85bd3734a3"} Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.744832 4741 scope.go:117] "RemoveContainer" containerID="ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.783499 4741 scope.go:117] "RemoveContainer" containerID="0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.802227 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5j4kk"] Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.811896 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5j4kk"] Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.813786 4741 scope.go:117] "RemoveContainer" containerID="54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.842962 4741 scope.go:117] "RemoveContainer" containerID="ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b" Sep 29 20:12:04 crc kubenswrapper[4741]: E0929 20:12:04.843720 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b\": container with ID starting with ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b not found: ID does not exist" containerID="ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.843897 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b"} err="failed to get container status \"ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b\": rpc error: code = NotFound desc = could not find container \"ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b\": container with ID starting with ae5858035d082f7c7a03f651971bad577533f06015420884cf6b62009da1c22b not found: ID does not exist" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.844137 4741 scope.go:117] "RemoveContainer" containerID="0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45" Sep 29 20:12:04 crc kubenswrapper[4741]: E0929 20:12:04.845033 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45\": container with ID starting with 0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45 not found: ID does not exist" containerID="0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.845085 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45"} err="failed to get container status \"0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45\": rpc error: code = NotFound desc = could not find container \"0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45\": container with ID starting with 0707b336058d2d139736fcb810a505b60b1df2fba9bc74cca98167da7c16cc45 not found: ID does not exist" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.845115 4741 scope.go:117] "RemoveContainer" containerID="54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a" Sep 29 20:12:04 crc kubenswrapper[4741]: E0929 20:12:04.845700 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a\": container with ID starting with 54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a not found: ID does not exist" containerID="54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a" Sep 29 20:12:04 crc kubenswrapper[4741]: I0929 20:12:04.845784 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a"} err="failed to get container status \"54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a\": rpc error: code = NotFound desc = could not find container \"54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a\": container with ID starting with 54b3d07758d63abc293ffd8ae103602a81b5d9809770e7d6f273bd7baef1125a not found: ID does not exist" Sep 29 20:12:05 crc kubenswrapper[4741]: I0929 20:12:05.102464 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" path="/var/lib/kubelet/pods/04020ceb-cea7-484f-bd11-ff4546937ced/volumes" Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.739600 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.740226 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.740287 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.741138 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.741238 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" gracePeriod=600 Sep 29 20:12:31 crc kubenswrapper[4741]: E0929 20:12:31.858917 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.972865 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" exitCode=0 Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.972916 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8"} Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.972947 4741 scope.go:117] "RemoveContainer" containerID="35e92b7732e29719fe07a9a393f0226f2393f56ecd021e5a4cd27179d979d479" Sep 29 20:12:31 crc kubenswrapper[4741]: I0929 20:12:31.973467 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:12:31 crc kubenswrapper[4741]: E0929 20:12:31.973701 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:12:44 crc kubenswrapper[4741]: I0929 20:12:44.086303 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:12:44 crc kubenswrapper[4741]: E0929 20:12:44.087274 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:12:58 crc kubenswrapper[4741]: I0929 20:12:58.085779 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:12:58 crc kubenswrapper[4741]: E0929 20:12:58.086839 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:13:11 crc kubenswrapper[4741]: I0929 20:13:11.086320 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:13:11 crc kubenswrapper[4741]: E0929 20:13:11.088495 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:13:24 crc kubenswrapper[4741]: I0929 20:13:24.086333 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:13:24 crc kubenswrapper[4741]: E0929 20:13:24.087288 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:13:36 crc kubenswrapper[4741]: I0929 20:13:36.087533 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:13:36 crc kubenswrapper[4741]: E0929 20:13:36.088434 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:13:49 crc kubenswrapper[4741]: I0929 20:13:49.094525 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:13:49 crc kubenswrapper[4741]: E0929 20:13:49.095575 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:14:03 crc kubenswrapper[4741]: I0929 20:14:03.086975 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:14:03 crc kubenswrapper[4741]: E0929 20:14:03.088102 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:14:14 crc kubenswrapper[4741]: I0929 20:14:14.085788 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:14:14 crc kubenswrapper[4741]: E0929 20:14:14.086534 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:14:28 crc kubenswrapper[4741]: I0929 20:14:28.085365 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:14:28 crc kubenswrapper[4741]: E0929 20:14:28.086087 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.779597 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cljf5"] Sep 29 20:14:32 crc kubenswrapper[4741]: E0929 20:14:32.780475 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="registry-server" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.780488 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="registry-server" Sep 29 20:14:32 crc kubenswrapper[4741]: E0929 20:14:32.780508 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="extract-utilities" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.780514 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="extract-utilities" Sep 29 20:14:32 crc kubenswrapper[4741]: E0929 20:14:32.780526 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="extract-content" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.780532 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="extract-content" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.780676 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="04020ceb-cea7-484f-bd11-ff4546937ced" containerName="registry-server" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.781726 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.792716 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cljf5"] Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.951021 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-catalog-content\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.951091 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqr55\" (UniqueName: \"kubernetes.io/projected/30c876e7-9f49-4c8a-9213-20914691b446-kube-api-access-mqr55\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:32 crc kubenswrapper[4741]: I0929 20:14:32.951187 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-utilities\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.052095 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-catalog-content\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.052157 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqr55\" (UniqueName: \"kubernetes.io/projected/30c876e7-9f49-4c8a-9213-20914691b446-kube-api-access-mqr55\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.052196 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-utilities\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.052609 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-utilities\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.052670 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-catalog-content\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.071360 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqr55\" (UniqueName: \"kubernetes.io/projected/30c876e7-9f49-4c8a-9213-20914691b446-kube-api-access-mqr55\") pod \"redhat-marketplace-cljf5\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.104021 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.586651 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cljf5"] Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.967597 4741 generic.go:334] "Generic (PLEG): container finished" podID="30c876e7-9f49-4c8a-9213-20914691b446" containerID="fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc" exitCode=0 Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.967637 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cljf5" event={"ID":"30c876e7-9f49-4c8a-9213-20914691b446","Type":"ContainerDied","Data":"fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc"} Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.967662 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cljf5" event={"ID":"30c876e7-9f49-4c8a-9213-20914691b446","Type":"ContainerStarted","Data":"4959b44d9b5f11ee029bc1a4b1a8e79da32449cd982dc7ee7521536f24661d3e"} Sep 29 20:14:33 crc kubenswrapper[4741]: I0929 20:14:33.969141 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:14:34 crc kubenswrapper[4741]: I0929 20:14:34.976596 4741 generic.go:334] "Generic (PLEG): container finished" podID="30c876e7-9f49-4c8a-9213-20914691b446" containerID="77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd" exitCode=0 Sep 29 20:14:34 crc kubenswrapper[4741]: I0929 20:14:34.976629 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cljf5" event={"ID":"30c876e7-9f49-4c8a-9213-20914691b446","Type":"ContainerDied","Data":"77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd"} Sep 29 20:14:35 crc kubenswrapper[4741]: I0929 20:14:35.984525 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cljf5" event={"ID":"30c876e7-9f49-4c8a-9213-20914691b446","Type":"ContainerStarted","Data":"4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50"} Sep 29 20:14:36 crc kubenswrapper[4741]: I0929 20:14:36.006019 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cljf5" podStartSLOduration=2.569027909 podStartE2EDuration="4.006003329s" podCreationTimestamp="2025-09-29 20:14:32 +0000 UTC" firstStartedPulling="2025-09-29 20:14:33.968950037 +0000 UTC m=+3915.616739369" lastFinishedPulling="2025-09-29 20:14:35.405925447 +0000 UTC m=+3917.053714789" observedRunningTime="2025-09-29 20:14:36.004189504 +0000 UTC m=+3917.651978836" watchObservedRunningTime="2025-09-29 20:14:36.006003329 +0000 UTC m=+3917.653792661" Sep 29 20:14:42 crc kubenswrapper[4741]: I0929 20:14:42.086967 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:14:42 crc kubenswrapper[4741]: E0929 20:14:42.088219 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:14:43 crc kubenswrapper[4741]: I0929 20:14:43.104432 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:43 crc kubenswrapper[4741]: I0929 20:14:43.105003 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:43 crc kubenswrapper[4741]: I0929 20:14:43.161696 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:44 crc kubenswrapper[4741]: I0929 20:14:44.080122 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:44 crc kubenswrapper[4741]: I0929 20:14:44.129964 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cljf5"] Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.056759 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cljf5" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="registry-server" containerID="cri-o://4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50" gracePeriod=2 Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.491752 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.644849 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqr55\" (UniqueName: \"kubernetes.io/projected/30c876e7-9f49-4c8a-9213-20914691b446-kube-api-access-mqr55\") pod \"30c876e7-9f49-4c8a-9213-20914691b446\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.644910 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-catalog-content\") pod \"30c876e7-9f49-4c8a-9213-20914691b446\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.644981 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-utilities\") pod \"30c876e7-9f49-4c8a-9213-20914691b446\" (UID: \"30c876e7-9f49-4c8a-9213-20914691b446\") " Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.646351 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-utilities" (OuterVolumeSpecName: "utilities") pod "30c876e7-9f49-4c8a-9213-20914691b446" (UID: "30c876e7-9f49-4c8a-9213-20914691b446"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.652646 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30c876e7-9f49-4c8a-9213-20914691b446-kube-api-access-mqr55" (OuterVolumeSpecName: "kube-api-access-mqr55") pod "30c876e7-9f49-4c8a-9213-20914691b446" (UID: "30c876e7-9f49-4c8a-9213-20914691b446"). InnerVolumeSpecName "kube-api-access-mqr55". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.662768 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30c876e7-9f49-4c8a-9213-20914691b446" (UID: "30c876e7-9f49-4c8a-9213-20914691b446"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.747237 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqr55\" (UniqueName: \"kubernetes.io/projected/30c876e7-9f49-4c8a-9213-20914691b446-kube-api-access-mqr55\") on node \"crc\" DevicePath \"\"" Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.747310 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:14:46 crc kubenswrapper[4741]: I0929 20:14:46.747339 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c876e7-9f49-4c8a-9213-20914691b446-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.091060 4741 generic.go:334] "Generic (PLEG): container finished" podID="30c876e7-9f49-4c8a-9213-20914691b446" containerID="4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50" exitCode=0 Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.091156 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cljf5" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.096266 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cljf5" event={"ID":"30c876e7-9f49-4c8a-9213-20914691b446","Type":"ContainerDied","Data":"4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50"} Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.096327 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cljf5" event={"ID":"30c876e7-9f49-4c8a-9213-20914691b446","Type":"ContainerDied","Data":"4959b44d9b5f11ee029bc1a4b1a8e79da32449cd982dc7ee7521536f24661d3e"} Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.096366 4741 scope.go:117] "RemoveContainer" containerID="4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.124940 4741 scope.go:117] "RemoveContainer" containerID="77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.144256 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cljf5"] Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.146610 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cljf5"] Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.148844 4741 scope.go:117] "RemoveContainer" containerID="fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.178815 4741 scope.go:117] "RemoveContainer" containerID="4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50" Sep 29 20:14:47 crc kubenswrapper[4741]: E0929 20:14:47.179437 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50\": container with ID starting with 4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50 not found: ID does not exist" containerID="4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.179477 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50"} err="failed to get container status \"4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50\": rpc error: code = NotFound desc = could not find container \"4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50\": container with ID starting with 4bd7944a79d615bbe17d366a8e936392cee006b2e502ca47c499d0f55d6e6a50 not found: ID does not exist" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.179506 4741 scope.go:117] "RemoveContainer" containerID="77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd" Sep 29 20:14:47 crc kubenswrapper[4741]: E0929 20:14:47.180013 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd\": container with ID starting with 77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd not found: ID does not exist" containerID="77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.180070 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd"} err="failed to get container status \"77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd\": rpc error: code = NotFound desc = could not find container \"77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd\": container with ID starting with 77e684574f4b6db2d17c9a57dc21d83242a7af747279655b29271a2a8d415fcd not found: ID does not exist" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.180108 4741 scope.go:117] "RemoveContainer" containerID="fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc" Sep 29 20:14:47 crc kubenswrapper[4741]: E0929 20:14:47.180644 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc\": container with ID starting with fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc not found: ID does not exist" containerID="fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc" Sep 29 20:14:47 crc kubenswrapper[4741]: I0929 20:14:47.180694 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc"} err="failed to get container status \"fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc\": rpc error: code = NotFound desc = could not find container \"fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc\": container with ID starting with fafc0fad8bdd468f8794cdfdc96012d2e6a8e6a581f84ff42fb4e44766f484cc not found: ID does not exist" Sep 29 20:14:49 crc kubenswrapper[4741]: I0929 20:14:49.099165 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30c876e7-9f49-4c8a-9213-20914691b446" path="/var/lib/kubelet/pods/30c876e7-9f49-4c8a-9213-20914691b446/volumes" Sep 29 20:14:57 crc kubenswrapper[4741]: I0929 20:14:57.087371 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:14:57 crc kubenswrapper[4741]: E0929 20:14:57.088126 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.166640 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s"] Sep 29 20:15:00 crc kubenswrapper[4741]: E0929 20:15:00.167288 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="registry-server" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.167304 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="registry-server" Sep 29 20:15:00 crc kubenswrapper[4741]: E0929 20:15:00.167325 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="extract-content" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.167333 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="extract-content" Sep 29 20:15:00 crc kubenswrapper[4741]: E0929 20:15:00.167359 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="extract-utilities" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.167370 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="extract-utilities" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.167567 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c876e7-9f49-4c8a-9213-20914691b446" containerName="registry-server" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.168319 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.172048 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.174607 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.182773 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s"] Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.248343 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fbd8d6-0243-4044-b907-8bf0448d58f2-config-volume\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.248456 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sjvf\" (UniqueName: \"kubernetes.io/projected/c7fbd8d6-0243-4044-b907-8bf0448d58f2-kube-api-access-2sjvf\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.248490 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7fbd8d6-0243-4044-b907-8bf0448d58f2-secret-volume\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.349238 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7fbd8d6-0243-4044-b907-8bf0448d58f2-secret-volume\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.349316 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fbd8d6-0243-4044-b907-8bf0448d58f2-config-volume\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.349368 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sjvf\" (UniqueName: \"kubernetes.io/projected/c7fbd8d6-0243-4044-b907-8bf0448d58f2-kube-api-access-2sjvf\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.351205 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fbd8d6-0243-4044-b907-8bf0448d58f2-config-volume\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.359076 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7fbd8d6-0243-4044-b907-8bf0448d58f2-secret-volume\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.364972 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sjvf\" (UniqueName: \"kubernetes.io/projected/c7fbd8d6-0243-4044-b907-8bf0448d58f2-kube-api-access-2sjvf\") pod \"collect-profiles-29319615-ftp9s\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.496353 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:00 crc kubenswrapper[4741]: I0929 20:15:00.703616 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s"] Sep 29 20:15:01 crc kubenswrapper[4741]: I0929 20:15:01.227858 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" event={"ID":"c7fbd8d6-0243-4044-b907-8bf0448d58f2","Type":"ContainerStarted","Data":"e423be380e915d310610a86192b1dd47582511ea8ce7d2d85805bd49737ac3c9"} Sep 29 20:15:01 crc kubenswrapper[4741]: I0929 20:15:01.227945 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" event={"ID":"c7fbd8d6-0243-4044-b907-8bf0448d58f2","Type":"ContainerStarted","Data":"cf82bd13847a88cacf9234df4a231d8e511c77209164204fb4e842275870ef2f"} Sep 29 20:15:01 crc kubenswrapper[4741]: I0929 20:15:01.248205 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" podStartSLOduration=1.248188124 podStartE2EDuration="1.248188124s" podCreationTimestamp="2025-09-29 20:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:15:01.247653837 +0000 UTC m=+3942.895443169" watchObservedRunningTime="2025-09-29 20:15:01.248188124 +0000 UTC m=+3942.895977456" Sep 29 20:15:02 crc kubenswrapper[4741]: I0929 20:15:02.238632 4741 generic.go:334] "Generic (PLEG): container finished" podID="c7fbd8d6-0243-4044-b907-8bf0448d58f2" containerID="e423be380e915d310610a86192b1dd47582511ea8ce7d2d85805bd49737ac3c9" exitCode=0 Sep 29 20:15:02 crc kubenswrapper[4741]: I0929 20:15:02.238706 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" event={"ID":"c7fbd8d6-0243-4044-b907-8bf0448d58f2","Type":"ContainerDied","Data":"e423be380e915d310610a86192b1dd47582511ea8ce7d2d85805bd49737ac3c9"} Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.515727 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.598352 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7fbd8d6-0243-4044-b907-8bf0448d58f2-secret-volume\") pod \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.598577 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sjvf\" (UniqueName: \"kubernetes.io/projected/c7fbd8d6-0243-4044-b907-8bf0448d58f2-kube-api-access-2sjvf\") pod \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.598621 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fbd8d6-0243-4044-b907-8bf0448d58f2-config-volume\") pod \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\" (UID: \"c7fbd8d6-0243-4044-b907-8bf0448d58f2\") " Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.599948 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7fbd8d6-0243-4044-b907-8bf0448d58f2-config-volume" (OuterVolumeSpecName: "config-volume") pod "c7fbd8d6-0243-4044-b907-8bf0448d58f2" (UID: "c7fbd8d6-0243-4044-b907-8bf0448d58f2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.605224 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7fbd8d6-0243-4044-b907-8bf0448d58f2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c7fbd8d6-0243-4044-b907-8bf0448d58f2" (UID: "c7fbd8d6-0243-4044-b907-8bf0448d58f2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.605618 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7fbd8d6-0243-4044-b907-8bf0448d58f2-kube-api-access-2sjvf" (OuterVolumeSpecName: "kube-api-access-2sjvf") pod "c7fbd8d6-0243-4044-b907-8bf0448d58f2" (UID: "c7fbd8d6-0243-4044-b907-8bf0448d58f2"). InnerVolumeSpecName "kube-api-access-2sjvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.701001 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7fbd8d6-0243-4044-b907-8bf0448d58f2-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.701059 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sjvf\" (UniqueName: \"kubernetes.io/projected/c7fbd8d6-0243-4044-b907-8bf0448d58f2-kube-api-access-2sjvf\") on node \"crc\" DevicePath \"\"" Sep 29 20:15:03 crc kubenswrapper[4741]: I0929 20:15:03.701080 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fbd8d6-0243-4044-b907-8bf0448d58f2-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:15:04 crc kubenswrapper[4741]: I0929 20:15:04.254994 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" event={"ID":"c7fbd8d6-0243-4044-b907-8bf0448d58f2","Type":"ContainerDied","Data":"cf82bd13847a88cacf9234df4a231d8e511c77209164204fb4e842275870ef2f"} Sep 29 20:15:04 crc kubenswrapper[4741]: I0929 20:15:04.255306 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf82bd13847a88cacf9234df4a231d8e511c77209164204fb4e842275870ef2f" Sep 29 20:15:04 crc kubenswrapper[4741]: I0929 20:15:04.255378 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s" Sep 29 20:15:04 crc kubenswrapper[4741]: I0929 20:15:04.319411 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595"] Sep 29 20:15:04 crc kubenswrapper[4741]: I0929 20:15:04.325930 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319570-lt595"] Sep 29 20:15:05 crc kubenswrapper[4741]: I0929 20:15:05.102216 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22434baf-68e0-4783-8835-27fa1e05b00d" path="/var/lib/kubelet/pods/22434baf-68e0-4783-8835-27fa1e05b00d/volumes" Sep 29 20:15:11 crc kubenswrapper[4741]: I0929 20:15:11.086195 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:15:11 crc kubenswrapper[4741]: E0929 20:15:11.086704 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:15:24 crc kubenswrapper[4741]: I0929 20:15:24.087088 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:15:24 crc kubenswrapper[4741]: E0929 20:15:24.088552 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:15:25 crc kubenswrapper[4741]: I0929 20:15:25.749796 4741 scope.go:117] "RemoveContainer" containerID="6d5598999cc4956e0cf131e4f176e34e0fe22baa68c7086cad3d29abcdeb7336" Sep 29 20:15:35 crc kubenswrapper[4741]: I0929 20:15:35.086129 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:15:35 crc kubenswrapper[4741]: E0929 20:15:35.086833 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:15:49 crc kubenswrapper[4741]: I0929 20:15:49.096755 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:15:49 crc kubenswrapper[4741]: E0929 20:15:49.097450 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:16:03 crc kubenswrapper[4741]: I0929 20:16:03.087176 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:16:03 crc kubenswrapper[4741]: E0929 20:16:03.088207 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:16:15 crc kubenswrapper[4741]: I0929 20:16:15.085940 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:16:15 crc kubenswrapper[4741]: E0929 20:16:15.087096 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:16:29 crc kubenswrapper[4741]: I0929 20:16:29.093894 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:16:29 crc kubenswrapper[4741]: E0929 20:16:29.094716 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:16:43 crc kubenswrapper[4741]: I0929 20:16:43.086157 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:16:43 crc kubenswrapper[4741]: E0929 20:16:43.086831 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:16:58 crc kubenswrapper[4741]: I0929 20:16:58.085993 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:16:58 crc kubenswrapper[4741]: E0929 20:16:58.087130 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:17:09 crc kubenswrapper[4741]: I0929 20:17:09.095689 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:17:09 crc kubenswrapper[4741]: E0929 20:17:09.096407 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:17:20 crc kubenswrapper[4741]: I0929 20:17:20.085935 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:17:20 crc kubenswrapper[4741]: E0929 20:17:20.086895 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:17:34 crc kubenswrapper[4741]: I0929 20:17:34.085991 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:17:34 crc kubenswrapper[4741]: I0929 20:17:34.625965 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"ccef0d12b1eb34848ee629bb3ec047bd5ae677186ad2661355829fe5d3833a26"} Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.757408 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c7qmx"] Sep 29 20:19:11 crc kubenswrapper[4741]: E0929 20:19:11.758498 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7fbd8d6-0243-4044-b907-8bf0448d58f2" containerName="collect-profiles" Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.758521 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7fbd8d6-0243-4044-b907-8bf0448d58f2" containerName="collect-profiles" Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.758751 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7fbd8d6-0243-4044-b907-8bf0448d58f2" containerName="collect-profiles" Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.761688 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.764401 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c7qmx"] Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.918258 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-catalog-content\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.918342 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-utilities\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:11 crc kubenswrapper[4741]: I0929 20:19:11.918408 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f6xz\" (UniqueName: \"kubernetes.io/projected/e2241801-1acc-4a52-9f50-1264c2e781d9-kube-api-access-6f6xz\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.019610 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-catalog-content\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.019991 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-utilities\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.020046 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f6xz\" (UniqueName: \"kubernetes.io/projected/e2241801-1acc-4a52-9f50-1264c2e781d9-kube-api-access-6f6xz\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.020162 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-catalog-content\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.020458 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-utilities\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.038236 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f6xz\" (UniqueName: \"kubernetes.io/projected/e2241801-1acc-4a52-9f50-1264c2e781d9-kube-api-access-6f6xz\") pod \"community-operators-c7qmx\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.108257 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:12 crc kubenswrapper[4741]: I0929 20:19:12.697467 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c7qmx"] Sep 29 20:19:13 crc kubenswrapper[4741]: I0929 20:19:13.476379 4741 generic.go:334] "Generic (PLEG): container finished" podID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerID="810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb" exitCode=0 Sep 29 20:19:13 crc kubenswrapper[4741]: I0929 20:19:13.476447 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c7qmx" event={"ID":"e2241801-1acc-4a52-9f50-1264c2e781d9","Type":"ContainerDied","Data":"810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb"} Sep 29 20:19:13 crc kubenswrapper[4741]: I0929 20:19:13.476782 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c7qmx" event={"ID":"e2241801-1acc-4a52-9f50-1264c2e781d9","Type":"ContainerStarted","Data":"2b11f2dfe151c4bcdb8ec4b071e70807c0214a354e0ed2523aec270491663ac9"} Sep 29 20:19:13 crc kubenswrapper[4741]: I0929 20:19:13.957795 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j8v6m"] Sep 29 20:19:13 crc kubenswrapper[4741]: I0929 20:19:13.959401 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:13 crc kubenswrapper[4741]: I0929 20:19:13.962947 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j8v6m"] Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.073875 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-utilities\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.073931 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4d7b\" (UniqueName: \"kubernetes.io/projected/f711b5ba-f3c8-4841-9b95-053dcb954019-kube-api-access-p4d7b\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.074023 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-catalog-content\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.175353 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-catalog-content\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.175432 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-utilities\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.175464 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4d7b\" (UniqueName: \"kubernetes.io/projected/f711b5ba-f3c8-4841-9b95-053dcb954019-kube-api-access-p4d7b\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.175987 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-catalog-content\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.176248 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-utilities\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.195252 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4d7b\" (UniqueName: \"kubernetes.io/projected/f711b5ba-f3c8-4841-9b95-053dcb954019-kube-api-access-p4d7b\") pod \"redhat-operators-j8v6m\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.277741 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.487815 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c7qmx" event={"ID":"e2241801-1acc-4a52-9f50-1264c2e781d9","Type":"ContainerStarted","Data":"5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28"} Sep 29 20:19:14 crc kubenswrapper[4741]: I0929 20:19:14.520040 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j8v6m"] Sep 29 20:19:14 crc kubenswrapper[4741]: W0929 20:19:14.522913 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf711b5ba_f3c8_4841_9b95_053dcb954019.slice/crio-d502c80e159891c0317b2becc4a447b94025fda96010fecdbd57e62c3eb9006b WatchSource:0}: Error finding container d502c80e159891c0317b2becc4a447b94025fda96010fecdbd57e62c3eb9006b: Status 404 returned error can't find the container with id d502c80e159891c0317b2becc4a447b94025fda96010fecdbd57e62c3eb9006b Sep 29 20:19:15 crc kubenswrapper[4741]: I0929 20:19:15.496190 4741 generic.go:334] "Generic (PLEG): container finished" podID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerID="e9dd0ee1373c6a051d6d595007dbd4e67bf1d020c118c038afba0e49d6b35a41" exitCode=0 Sep 29 20:19:15 crc kubenswrapper[4741]: I0929 20:19:15.496262 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8v6m" event={"ID":"f711b5ba-f3c8-4841-9b95-053dcb954019","Type":"ContainerDied","Data":"e9dd0ee1373c6a051d6d595007dbd4e67bf1d020c118c038afba0e49d6b35a41"} Sep 29 20:19:15 crc kubenswrapper[4741]: I0929 20:19:15.496590 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8v6m" event={"ID":"f711b5ba-f3c8-4841-9b95-053dcb954019","Type":"ContainerStarted","Data":"d502c80e159891c0317b2becc4a447b94025fda96010fecdbd57e62c3eb9006b"} Sep 29 20:19:15 crc kubenswrapper[4741]: I0929 20:19:15.500066 4741 generic.go:334] "Generic (PLEG): container finished" podID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerID="5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28" exitCode=0 Sep 29 20:19:15 crc kubenswrapper[4741]: I0929 20:19:15.500095 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c7qmx" event={"ID":"e2241801-1acc-4a52-9f50-1264c2e781d9","Type":"ContainerDied","Data":"5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28"} Sep 29 20:19:16 crc kubenswrapper[4741]: I0929 20:19:16.511196 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c7qmx" event={"ID":"e2241801-1acc-4a52-9f50-1264c2e781d9","Type":"ContainerStarted","Data":"a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c"} Sep 29 20:19:16 crc kubenswrapper[4741]: I0929 20:19:16.512993 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8v6m" event={"ID":"f711b5ba-f3c8-4841-9b95-053dcb954019","Type":"ContainerStarted","Data":"16ba7158e98fe2ecdb48519ff241c44f3ba291a6264222ba6f2761594c93b316"} Sep 29 20:19:16 crc kubenswrapper[4741]: I0929 20:19:16.527617 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c7qmx" podStartSLOduration=3.070255325 podStartE2EDuration="5.527598447s" podCreationTimestamp="2025-09-29 20:19:11 +0000 UTC" firstStartedPulling="2025-09-29 20:19:13.47765481 +0000 UTC m=+4195.125444142" lastFinishedPulling="2025-09-29 20:19:15.934997922 +0000 UTC m=+4197.582787264" observedRunningTime="2025-09-29 20:19:16.525650176 +0000 UTC m=+4198.173439538" watchObservedRunningTime="2025-09-29 20:19:16.527598447 +0000 UTC m=+4198.175387779" Sep 29 20:19:17 crc kubenswrapper[4741]: I0929 20:19:17.520757 4741 generic.go:334] "Generic (PLEG): container finished" podID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerID="16ba7158e98fe2ecdb48519ff241c44f3ba291a6264222ba6f2761594c93b316" exitCode=0 Sep 29 20:19:17 crc kubenswrapper[4741]: I0929 20:19:17.520826 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8v6m" event={"ID":"f711b5ba-f3c8-4841-9b95-053dcb954019","Type":"ContainerDied","Data":"16ba7158e98fe2ecdb48519ff241c44f3ba291a6264222ba6f2761594c93b316"} Sep 29 20:19:18 crc kubenswrapper[4741]: I0929 20:19:18.528760 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8v6m" event={"ID":"f711b5ba-f3c8-4841-9b95-053dcb954019","Type":"ContainerStarted","Data":"690c3bea4338ff00584ded7b577692036a400885dae30dbf775c43f936d0eefb"} Sep 29 20:19:18 crc kubenswrapper[4741]: I0929 20:19:18.545487 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j8v6m" podStartSLOduration=3.139732447 podStartE2EDuration="5.545468583s" podCreationTimestamp="2025-09-29 20:19:13 +0000 UTC" firstStartedPulling="2025-09-29 20:19:15.497942831 +0000 UTC m=+4197.145732163" lastFinishedPulling="2025-09-29 20:19:17.903678967 +0000 UTC m=+4199.551468299" observedRunningTime="2025-09-29 20:19:18.544100789 +0000 UTC m=+4200.191890121" watchObservedRunningTime="2025-09-29 20:19:18.545468583 +0000 UTC m=+4200.193257915" Sep 29 20:19:22 crc kubenswrapper[4741]: I0929 20:19:22.109440 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:22 crc kubenswrapper[4741]: I0929 20:19:22.111506 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:22 crc kubenswrapper[4741]: I0929 20:19:22.158555 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:22 crc kubenswrapper[4741]: I0929 20:19:22.593857 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:22 crc kubenswrapper[4741]: I0929 20:19:22.637665 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c7qmx"] Sep 29 20:19:24 crc kubenswrapper[4741]: I0929 20:19:24.278891 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:24 crc kubenswrapper[4741]: I0929 20:19:24.279248 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:24 crc kubenswrapper[4741]: I0929 20:19:24.319001 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:24 crc kubenswrapper[4741]: I0929 20:19:24.571621 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c7qmx" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="registry-server" containerID="cri-o://a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c" gracePeriod=2 Sep 29 20:19:24 crc kubenswrapper[4741]: I0929 20:19:24.618340 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:24 crc kubenswrapper[4741]: I0929 20:19:24.946166 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j8v6m"] Sep 29 20:19:24 crc kubenswrapper[4741]: I0929 20:19:24.983108 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.122790 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-catalog-content\") pod \"e2241801-1acc-4a52-9f50-1264c2e781d9\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.122848 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f6xz\" (UniqueName: \"kubernetes.io/projected/e2241801-1acc-4a52-9f50-1264c2e781d9-kube-api-access-6f6xz\") pod \"e2241801-1acc-4a52-9f50-1264c2e781d9\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.122991 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-utilities\") pod \"e2241801-1acc-4a52-9f50-1264c2e781d9\" (UID: \"e2241801-1acc-4a52-9f50-1264c2e781d9\") " Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.124262 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-utilities" (OuterVolumeSpecName: "utilities") pod "e2241801-1acc-4a52-9f50-1264c2e781d9" (UID: "e2241801-1acc-4a52-9f50-1264c2e781d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.129595 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2241801-1acc-4a52-9f50-1264c2e781d9-kube-api-access-6f6xz" (OuterVolumeSpecName: "kube-api-access-6f6xz") pod "e2241801-1acc-4a52-9f50-1264c2e781d9" (UID: "e2241801-1acc-4a52-9f50-1264c2e781d9"). InnerVolumeSpecName "kube-api-access-6f6xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.224608 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.224640 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f6xz\" (UniqueName: \"kubernetes.io/projected/e2241801-1acc-4a52-9f50-1264c2e781d9-kube-api-access-6f6xz\") on node \"crc\" DevicePath \"\"" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.492072 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2241801-1acc-4a52-9f50-1264c2e781d9" (UID: "e2241801-1acc-4a52-9f50-1264c2e781d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.529041 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2241801-1acc-4a52-9f50-1264c2e781d9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.581434 4741 generic.go:334] "Generic (PLEG): container finished" podID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerID="a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c" exitCode=0 Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.581573 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c7qmx" event={"ID":"e2241801-1acc-4a52-9f50-1264c2e781d9","Type":"ContainerDied","Data":"a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c"} Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.581628 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c7qmx" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.581639 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c7qmx" event={"ID":"e2241801-1acc-4a52-9f50-1264c2e781d9","Type":"ContainerDied","Data":"2b11f2dfe151c4bcdb8ec4b071e70807c0214a354e0ed2523aec270491663ac9"} Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.581827 4741 scope.go:117] "RemoveContainer" containerID="a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.605609 4741 scope.go:117] "RemoveContainer" containerID="5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.618751 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c7qmx"] Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.636317 4741 scope.go:117] "RemoveContainer" containerID="810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.644843 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c7qmx"] Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.668720 4741 scope.go:117] "RemoveContainer" containerID="a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c" Sep 29 20:19:25 crc kubenswrapper[4741]: E0929 20:19:25.669222 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c\": container with ID starting with a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c not found: ID does not exist" containerID="a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.669264 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c"} err="failed to get container status \"a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c\": rpc error: code = NotFound desc = could not find container \"a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c\": container with ID starting with a416e6e9559858746eab4bdeb27341a9a2832ed829535969772af42bc447361c not found: ID does not exist" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.669289 4741 scope.go:117] "RemoveContainer" containerID="5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28" Sep 29 20:19:25 crc kubenswrapper[4741]: E0929 20:19:25.669791 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28\": container with ID starting with 5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28 not found: ID does not exist" containerID="5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.669812 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28"} err="failed to get container status \"5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28\": rpc error: code = NotFound desc = could not find container \"5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28\": container with ID starting with 5fa3f13c8dda8348dc70bffa16ec62e47d0514758c33d6a6e8a347f4f7daac28 not found: ID does not exist" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.669825 4741 scope.go:117] "RemoveContainer" containerID="810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb" Sep 29 20:19:25 crc kubenswrapper[4741]: E0929 20:19:25.670135 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb\": container with ID starting with 810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb not found: ID does not exist" containerID="810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb" Sep 29 20:19:25 crc kubenswrapper[4741]: I0929 20:19:25.670182 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb"} err="failed to get container status \"810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb\": rpc error: code = NotFound desc = could not find container \"810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb\": container with ID starting with 810f69ce0320e36f99b778e1b37c6726e47fcc78cee4fd93f87a091838282feb not found: ID does not exist" Sep 29 20:19:26 crc kubenswrapper[4741]: I0929 20:19:26.587835 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j8v6m" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="registry-server" containerID="cri-o://690c3bea4338ff00584ded7b577692036a400885dae30dbf775c43f936d0eefb" gracePeriod=2 Sep 29 20:19:27 crc kubenswrapper[4741]: I0929 20:19:27.093267 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" path="/var/lib/kubelet/pods/e2241801-1acc-4a52-9f50-1264c2e781d9/volumes" Sep 29 20:19:27 crc kubenswrapper[4741]: I0929 20:19:27.599909 4741 generic.go:334] "Generic (PLEG): container finished" podID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerID="690c3bea4338ff00584ded7b577692036a400885dae30dbf775c43f936d0eefb" exitCode=0 Sep 29 20:19:27 crc kubenswrapper[4741]: I0929 20:19:27.600011 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8v6m" event={"ID":"f711b5ba-f3c8-4841-9b95-053dcb954019","Type":"ContainerDied","Data":"690c3bea4338ff00584ded7b577692036a400885dae30dbf775c43f936d0eefb"} Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.367248 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.489671 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-utilities\") pod \"f711b5ba-f3c8-4841-9b95-053dcb954019\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.490247 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-catalog-content\") pod \"f711b5ba-f3c8-4841-9b95-053dcb954019\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.490335 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4d7b\" (UniqueName: \"kubernetes.io/projected/f711b5ba-f3c8-4841-9b95-053dcb954019-kube-api-access-p4d7b\") pod \"f711b5ba-f3c8-4841-9b95-053dcb954019\" (UID: \"f711b5ba-f3c8-4841-9b95-053dcb954019\") " Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.491812 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-utilities" (OuterVolumeSpecName: "utilities") pod "f711b5ba-f3c8-4841-9b95-053dcb954019" (UID: "f711b5ba-f3c8-4841-9b95-053dcb954019"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.495717 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f711b5ba-f3c8-4841-9b95-053dcb954019-kube-api-access-p4d7b" (OuterVolumeSpecName: "kube-api-access-p4d7b") pod "f711b5ba-f3c8-4841-9b95-053dcb954019" (UID: "f711b5ba-f3c8-4841-9b95-053dcb954019"). InnerVolumeSpecName "kube-api-access-p4d7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.578247 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f711b5ba-f3c8-4841-9b95-053dcb954019" (UID: "f711b5ba-f3c8-4841-9b95-053dcb954019"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.591582 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4d7b\" (UniqueName: \"kubernetes.io/projected/f711b5ba-f3c8-4841-9b95-053dcb954019-kube-api-access-p4d7b\") on node \"crc\" DevicePath \"\"" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.591606 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.591615 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f711b5ba-f3c8-4841-9b95-053dcb954019-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.608688 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8v6m" event={"ID":"f711b5ba-f3c8-4841-9b95-053dcb954019","Type":"ContainerDied","Data":"d502c80e159891c0317b2becc4a447b94025fda96010fecdbd57e62c3eb9006b"} Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.608740 4741 scope.go:117] "RemoveContainer" containerID="690c3bea4338ff00584ded7b577692036a400885dae30dbf775c43f936d0eefb" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.608935 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8v6m" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.626805 4741 scope.go:117] "RemoveContainer" containerID="16ba7158e98fe2ecdb48519ff241c44f3ba291a6264222ba6f2761594c93b316" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.644013 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j8v6m"] Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.647191 4741 scope.go:117] "RemoveContainer" containerID="e9dd0ee1373c6a051d6d595007dbd4e67bf1d020c118c038afba0e49d6b35a41" Sep 29 20:19:28 crc kubenswrapper[4741]: I0929 20:19:28.649554 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j8v6m"] Sep 29 20:19:29 crc kubenswrapper[4741]: I0929 20:19:29.111441 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" path="/var/lib/kubelet/pods/f711b5ba-f3c8-4841-9b95-053dcb954019/volumes" Sep 29 20:20:01 crc kubenswrapper[4741]: I0929 20:20:01.739772 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:20:01 crc kubenswrapper[4741]: I0929 20:20:01.740423 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:20:31 crc kubenswrapper[4741]: I0929 20:20:31.739061 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:20:31 crc kubenswrapper[4741]: I0929 20:20:31.739783 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:21:01 crc kubenswrapper[4741]: I0929 20:21:01.739065 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:21:01 crc kubenswrapper[4741]: I0929 20:21:01.739903 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:21:01 crc kubenswrapper[4741]: I0929 20:21:01.739979 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:21:01 crc kubenswrapper[4741]: I0929 20:21:01.740842 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ccef0d12b1eb34848ee629bb3ec047bd5ae677186ad2661355829fe5d3833a26"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:21:01 crc kubenswrapper[4741]: I0929 20:21:01.740943 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://ccef0d12b1eb34848ee629bb3ec047bd5ae677186ad2661355829fe5d3833a26" gracePeriod=600 Sep 29 20:21:02 crc kubenswrapper[4741]: I0929 20:21:02.427598 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="ccef0d12b1eb34848ee629bb3ec047bd5ae677186ad2661355829fe5d3833a26" exitCode=0 Sep 29 20:21:02 crc kubenswrapper[4741]: I0929 20:21:02.427660 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"ccef0d12b1eb34848ee629bb3ec047bd5ae677186ad2661355829fe5d3833a26"} Sep 29 20:21:02 crc kubenswrapper[4741]: I0929 20:21:02.428150 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2"} Sep 29 20:21:02 crc kubenswrapper[4741]: I0929 20:21:02.428174 4741 scope.go:117] "RemoveContainer" containerID="acdd57c65cfc4fd0f3958fa2c7a23c90033ada91fbd529c0c716037a64a4dfa8" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.316600 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kbkp5"] Sep 29 20:23:01 crc kubenswrapper[4741]: E0929 20:23:01.317501 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="extract-content" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317519 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="extract-content" Sep 29 20:23:01 crc kubenswrapper[4741]: E0929 20:23:01.317538 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="registry-server" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317547 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="registry-server" Sep 29 20:23:01 crc kubenswrapper[4741]: E0929 20:23:01.317566 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="extract-utilities" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317573 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="extract-utilities" Sep 29 20:23:01 crc kubenswrapper[4741]: E0929 20:23:01.317647 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="extract-content" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317658 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="extract-content" Sep 29 20:23:01 crc kubenswrapper[4741]: E0929 20:23:01.317670 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="extract-utilities" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317678 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="extract-utilities" Sep 29 20:23:01 crc kubenswrapper[4741]: E0929 20:23:01.317689 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="registry-server" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317720 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="registry-server" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317927 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2241801-1acc-4a52-9f50-1264c2e781d9" containerName="registry-server" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.317973 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f711b5ba-f3c8-4841-9b95-053dcb954019" containerName="registry-server" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.319573 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.334209 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbkp5"] Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.506575 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-catalog-content\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.506950 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t7jk\" (UniqueName: \"kubernetes.io/projected/9de2ae30-b9b5-4d18-8e00-f941b7850e12-kube-api-access-9t7jk\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.507005 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-utilities\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.608803 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t7jk\" (UniqueName: \"kubernetes.io/projected/9de2ae30-b9b5-4d18-8e00-f941b7850e12-kube-api-access-9t7jk\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.608921 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-utilities\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.608989 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-catalog-content\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.609506 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-utilities\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.609670 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-catalog-content\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.640606 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t7jk\" (UniqueName: \"kubernetes.io/projected/9de2ae30-b9b5-4d18-8e00-f941b7850e12-kube-api-access-9t7jk\") pod \"certified-operators-kbkp5\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:01 crc kubenswrapper[4741]: I0929 20:23:01.939588 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:02 crc kubenswrapper[4741]: I0929 20:23:02.427170 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbkp5"] Sep 29 20:23:02 crc kubenswrapper[4741]: I0929 20:23:02.480702 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbkp5" event={"ID":"9de2ae30-b9b5-4d18-8e00-f941b7850e12","Type":"ContainerStarted","Data":"9802efe303d6ae489942f711932e821cf03d2f54da86b8e31888b49f640540d2"} Sep 29 20:23:02 crc kubenswrapper[4741]: I0929 20:23:02.982845 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-6mdn7"] Sep 29 20:23:02 crc kubenswrapper[4741]: I0929 20:23:02.988127 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-6mdn7"] Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.101882 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f516a891-abe1-4143-9492-7bd23d5e74a9" path="/var/lib/kubelet/pods/f516a891-abe1-4143-9492-7bd23d5e74a9/volumes" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.157056 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-r4hlp"] Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.159278 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.161053 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.162628 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.162645 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.163123 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-r4hlp"] Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.163212 4741 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-q2245" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.235539 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66d17356-9220-4ee1-86e4-c044992df64b-crc-storage\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.235626 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2dzf\" (UniqueName: \"kubernetes.io/projected/66d17356-9220-4ee1-86e4-c044992df64b-kube-api-access-l2dzf\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.235648 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66d17356-9220-4ee1-86e4-c044992df64b-node-mnt\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.336783 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66d17356-9220-4ee1-86e4-c044992df64b-crc-storage\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.336986 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2dzf\" (UniqueName: \"kubernetes.io/projected/66d17356-9220-4ee1-86e4-c044992df64b-kube-api-access-l2dzf\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.337041 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66d17356-9220-4ee1-86e4-c044992df64b-node-mnt\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.337508 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66d17356-9220-4ee1-86e4-c044992df64b-crc-storage\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.337585 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66d17356-9220-4ee1-86e4-c044992df64b-node-mnt\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.355985 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2dzf\" (UniqueName: \"kubernetes.io/projected/66d17356-9220-4ee1-86e4-c044992df64b-kube-api-access-l2dzf\") pod \"crc-storage-crc-r4hlp\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.484344 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.499692 4741 generic.go:334] "Generic (PLEG): container finished" podID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerID="28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a" exitCode=0 Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.499747 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbkp5" event={"ID":"9de2ae30-b9b5-4d18-8e00-f941b7850e12","Type":"ContainerDied","Data":"28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a"} Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.505716 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:23:03 crc kubenswrapper[4741]: I0929 20:23:03.723233 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-r4hlp"] Sep 29 20:23:03 crc kubenswrapper[4741]: W0929 20:23:03.725523 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66d17356_9220_4ee1_86e4_c044992df64b.slice/crio-6f4fc1d98cbc0a8076d2524e7bf6db4c424a91f6f67400ee4fb0f584bb7b6f66 WatchSource:0}: Error finding container 6f4fc1d98cbc0a8076d2524e7bf6db4c424a91f6f67400ee4fb0f584bb7b6f66: Status 404 returned error can't find the container with id 6f4fc1d98cbc0a8076d2524e7bf6db4c424a91f6f67400ee4fb0f584bb7b6f66 Sep 29 20:23:04 crc kubenswrapper[4741]: I0929 20:23:04.509592 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r4hlp" event={"ID":"66d17356-9220-4ee1-86e4-c044992df64b","Type":"ContainerStarted","Data":"32aae837832bdd03c822cf2fb7c9f7642fcb0bd2ac8435ebb3794c0b27312e5e"} Sep 29 20:23:04 crc kubenswrapper[4741]: I0929 20:23:04.510140 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r4hlp" event={"ID":"66d17356-9220-4ee1-86e4-c044992df64b","Type":"ContainerStarted","Data":"6f4fc1d98cbc0a8076d2524e7bf6db4c424a91f6f67400ee4fb0f584bb7b6f66"} Sep 29 20:23:04 crc kubenswrapper[4741]: I0929 20:23:04.511665 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbkp5" event={"ID":"9de2ae30-b9b5-4d18-8e00-f941b7850e12","Type":"ContainerStarted","Data":"8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5"} Sep 29 20:23:04 crc kubenswrapper[4741]: I0929 20:23:04.528414 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-r4hlp" podStartSLOduration=1.027119877 podStartE2EDuration="1.528376142s" podCreationTimestamp="2025-09-29 20:23:03 +0000 UTC" firstStartedPulling="2025-09-29 20:23:03.728022477 +0000 UTC m=+4425.375811819" lastFinishedPulling="2025-09-29 20:23:04.229278742 +0000 UTC m=+4425.877068084" observedRunningTime="2025-09-29 20:23:04.525841733 +0000 UTC m=+4426.173631085" watchObservedRunningTime="2025-09-29 20:23:04.528376142 +0000 UTC m=+4426.176165474" Sep 29 20:23:05 crc kubenswrapper[4741]: I0929 20:23:05.525974 4741 generic.go:334] "Generic (PLEG): container finished" podID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerID="8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5" exitCode=0 Sep 29 20:23:05 crc kubenswrapper[4741]: I0929 20:23:05.526055 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbkp5" event={"ID":"9de2ae30-b9b5-4d18-8e00-f941b7850e12","Type":"ContainerDied","Data":"8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5"} Sep 29 20:23:05 crc kubenswrapper[4741]: I0929 20:23:05.528815 4741 generic.go:334] "Generic (PLEG): container finished" podID="66d17356-9220-4ee1-86e4-c044992df64b" containerID="32aae837832bdd03c822cf2fb7c9f7642fcb0bd2ac8435ebb3794c0b27312e5e" exitCode=0 Sep 29 20:23:05 crc kubenswrapper[4741]: I0929 20:23:05.528844 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r4hlp" event={"ID":"66d17356-9220-4ee1-86e4-c044992df64b","Type":"ContainerDied","Data":"32aae837832bdd03c822cf2fb7c9f7642fcb0bd2ac8435ebb3794c0b27312e5e"} Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.541682 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbkp5" event={"ID":"9de2ae30-b9b5-4d18-8e00-f941b7850e12","Type":"ContainerStarted","Data":"34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f"} Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.809681 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.831247 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kbkp5" podStartSLOduration=3.39034033 podStartE2EDuration="5.831231334s" podCreationTimestamp="2025-09-29 20:23:01 +0000 UTC" firstStartedPulling="2025-09-29 20:23:03.505454459 +0000 UTC m=+4425.153243781" lastFinishedPulling="2025-09-29 20:23:05.946345433 +0000 UTC m=+4427.594134785" observedRunningTime="2025-09-29 20:23:06.56692777 +0000 UTC m=+4428.214717152" watchObservedRunningTime="2025-09-29 20:23:06.831231334 +0000 UTC m=+4428.479020666" Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.985750 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66d17356-9220-4ee1-86e4-c044992df64b-node-mnt\") pod \"66d17356-9220-4ee1-86e4-c044992df64b\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.985863 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66d17356-9220-4ee1-86e4-c044992df64b-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "66d17356-9220-4ee1-86e4-c044992df64b" (UID: "66d17356-9220-4ee1-86e4-c044992df64b"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.985931 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66d17356-9220-4ee1-86e4-c044992df64b-crc-storage\") pod \"66d17356-9220-4ee1-86e4-c044992df64b\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.985971 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2dzf\" (UniqueName: \"kubernetes.io/projected/66d17356-9220-4ee1-86e4-c044992df64b-kube-api-access-l2dzf\") pod \"66d17356-9220-4ee1-86e4-c044992df64b\" (UID: \"66d17356-9220-4ee1-86e4-c044992df64b\") " Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.986211 4741 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66d17356-9220-4ee1-86e4-c044992df64b-node-mnt\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:06 crc kubenswrapper[4741]: I0929 20:23:06.990559 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d17356-9220-4ee1-86e4-c044992df64b-kube-api-access-l2dzf" (OuterVolumeSpecName: "kube-api-access-l2dzf") pod "66d17356-9220-4ee1-86e4-c044992df64b" (UID: "66d17356-9220-4ee1-86e4-c044992df64b"). InnerVolumeSpecName "kube-api-access-l2dzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:23:07 crc kubenswrapper[4741]: I0929 20:23:07.006066 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d17356-9220-4ee1-86e4-c044992df64b-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "66d17356-9220-4ee1-86e4-c044992df64b" (UID: "66d17356-9220-4ee1-86e4-c044992df64b"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:23:07 crc kubenswrapper[4741]: I0929 20:23:07.087259 4741 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66d17356-9220-4ee1-86e4-c044992df64b-crc-storage\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:07 crc kubenswrapper[4741]: I0929 20:23:07.087293 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2dzf\" (UniqueName: \"kubernetes.io/projected/66d17356-9220-4ee1-86e4-c044992df64b-kube-api-access-l2dzf\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:07 crc kubenswrapper[4741]: I0929 20:23:07.549479 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r4hlp" event={"ID":"66d17356-9220-4ee1-86e4-c044992df64b","Type":"ContainerDied","Data":"6f4fc1d98cbc0a8076d2524e7bf6db4c424a91f6f67400ee4fb0f584bb7b6f66"} Sep 29 20:23:07 crc kubenswrapper[4741]: I0929 20:23:07.549530 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r4hlp" Sep 29 20:23:07 crc kubenswrapper[4741]: I0929 20:23:07.549537 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f4fc1d98cbc0a8076d2524e7bf6db4c424a91f6f67400ee4fb0f584bb7b6f66" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.268310 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-r4hlp"] Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.279476 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-r4hlp"] Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.440929 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-s4n9k"] Sep 29 20:23:09 crc kubenswrapper[4741]: E0929 20:23:09.441758 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d17356-9220-4ee1-86e4-c044992df64b" containerName="storage" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.441791 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d17356-9220-4ee1-86e4-c044992df64b" containerName="storage" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.442098 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d17356-9220-4ee1-86e4-c044992df64b" containerName="storage" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.442890 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.445446 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.445555 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.445643 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.445870 4741 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-q2245" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.450197 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-s4n9k"] Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.624410 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/226db4b2-8bb4-44de-a918-4d933b4c31ef-crc-storage\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.624469 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/226db4b2-8bb4-44de-a918-4d933b4c31ef-node-mnt\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.624490 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm64p\" (UniqueName: \"kubernetes.io/projected/226db4b2-8bb4-44de-a918-4d933b4c31ef-kube-api-access-mm64p\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.726108 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/226db4b2-8bb4-44de-a918-4d933b4c31ef-crc-storage\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.726185 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/226db4b2-8bb4-44de-a918-4d933b4c31ef-node-mnt\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.726217 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm64p\" (UniqueName: \"kubernetes.io/projected/226db4b2-8bb4-44de-a918-4d933b4c31ef-kube-api-access-mm64p\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.726523 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/226db4b2-8bb4-44de-a918-4d933b4c31ef-node-mnt\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.727152 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/226db4b2-8bb4-44de-a918-4d933b4c31ef-crc-storage\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.751480 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm64p\" (UniqueName: \"kubernetes.io/projected/226db4b2-8bb4-44de-a918-4d933b4c31ef-kube-api-access-mm64p\") pod \"crc-storage-crc-s4n9k\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:09 crc kubenswrapper[4741]: I0929 20:23:09.768535 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:10 crc kubenswrapper[4741]: I0929 20:23:10.219833 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-s4n9k"] Sep 29 20:23:10 crc kubenswrapper[4741]: I0929 20:23:10.571266 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-s4n9k" event={"ID":"226db4b2-8bb4-44de-a918-4d933b4c31ef","Type":"ContainerStarted","Data":"4403bdc1ee19130ad4c77743f2c85904425ba2b2a509423e0dcf1a9971462306"} Sep 29 20:23:11 crc kubenswrapper[4741]: I0929 20:23:11.099109 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d17356-9220-4ee1-86e4-c044992df64b" path="/var/lib/kubelet/pods/66d17356-9220-4ee1-86e4-c044992df64b/volumes" Sep 29 20:23:11 crc kubenswrapper[4741]: I0929 20:23:11.581439 4741 generic.go:334] "Generic (PLEG): container finished" podID="226db4b2-8bb4-44de-a918-4d933b4c31ef" containerID="1a4e5451595468cafc69f332371fbcba8a60e8817b9086dcd9988be84069d285" exitCode=0 Sep 29 20:23:11 crc kubenswrapper[4741]: I0929 20:23:11.581552 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-s4n9k" event={"ID":"226db4b2-8bb4-44de-a918-4d933b4c31ef","Type":"ContainerDied","Data":"1a4e5451595468cafc69f332371fbcba8a60e8817b9086dcd9988be84069d285"} Sep 29 20:23:11 crc kubenswrapper[4741]: I0929 20:23:11.939957 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:11 crc kubenswrapper[4741]: I0929 20:23:11.941467 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:11 crc kubenswrapper[4741]: I0929 20:23:11.987842 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:12 crc kubenswrapper[4741]: I0929 20:23:12.851112 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:12 crc kubenswrapper[4741]: I0929 20:23:12.907693 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbkp5"] Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.033547 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.177306 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm64p\" (UniqueName: \"kubernetes.io/projected/226db4b2-8bb4-44de-a918-4d933b4c31ef-kube-api-access-mm64p\") pod \"226db4b2-8bb4-44de-a918-4d933b4c31ef\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.177482 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/226db4b2-8bb4-44de-a918-4d933b4c31ef-crc-storage\") pod \"226db4b2-8bb4-44de-a918-4d933b4c31ef\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.177563 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/226db4b2-8bb4-44de-a918-4d933b4c31ef-node-mnt\") pod \"226db4b2-8bb4-44de-a918-4d933b4c31ef\" (UID: \"226db4b2-8bb4-44de-a918-4d933b4c31ef\") " Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.177863 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/226db4b2-8bb4-44de-a918-4d933b4c31ef-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "226db4b2-8bb4-44de-a918-4d933b4c31ef" (UID: "226db4b2-8bb4-44de-a918-4d933b4c31ef"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.178452 4741 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/226db4b2-8bb4-44de-a918-4d933b4c31ef-node-mnt\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.182352 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/226db4b2-8bb4-44de-a918-4d933b4c31ef-kube-api-access-mm64p" (OuterVolumeSpecName: "kube-api-access-mm64p") pod "226db4b2-8bb4-44de-a918-4d933b4c31ef" (UID: "226db4b2-8bb4-44de-a918-4d933b4c31ef"). InnerVolumeSpecName "kube-api-access-mm64p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.211515 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/226db4b2-8bb4-44de-a918-4d933b4c31ef-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "226db4b2-8bb4-44de-a918-4d933b4c31ef" (UID: "226db4b2-8bb4-44de-a918-4d933b4c31ef"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.279836 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm64p\" (UniqueName: \"kubernetes.io/projected/226db4b2-8bb4-44de-a918-4d933b4c31ef-kube-api-access-mm64p\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.279900 4741 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/226db4b2-8bb4-44de-a918-4d933b4c31ef-crc-storage\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.599151 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-s4n9k" event={"ID":"226db4b2-8bb4-44de-a918-4d933b4c31ef","Type":"ContainerDied","Data":"4403bdc1ee19130ad4c77743f2c85904425ba2b2a509423e0dcf1a9971462306"} Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.599235 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4403bdc1ee19130ad4c77743f2c85904425ba2b2a509423e0dcf1a9971462306" Sep 29 20:23:13 crc kubenswrapper[4741]: I0929 20:23:13.599237 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4n9k" Sep 29 20:23:14 crc kubenswrapper[4741]: I0929 20:23:14.606834 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kbkp5" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="registry-server" containerID="cri-o://34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f" gracePeriod=2 Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.175301 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.310998 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-utilities\") pod \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.311085 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t7jk\" (UniqueName: \"kubernetes.io/projected/9de2ae30-b9b5-4d18-8e00-f941b7850e12-kube-api-access-9t7jk\") pod \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.311132 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-catalog-content\") pod \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\" (UID: \"9de2ae30-b9b5-4d18-8e00-f941b7850e12\") " Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.311970 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-utilities" (OuterVolumeSpecName: "utilities") pod "9de2ae30-b9b5-4d18-8e00-f941b7850e12" (UID: "9de2ae30-b9b5-4d18-8e00-f941b7850e12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.317809 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.326269 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9de2ae30-b9b5-4d18-8e00-f941b7850e12-kube-api-access-9t7jk" (OuterVolumeSpecName: "kube-api-access-9t7jk") pod "9de2ae30-b9b5-4d18-8e00-f941b7850e12" (UID: "9de2ae30-b9b5-4d18-8e00-f941b7850e12"). InnerVolumeSpecName "kube-api-access-9t7jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.360848 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9de2ae30-b9b5-4d18-8e00-f941b7850e12" (UID: "9de2ae30-b9b5-4d18-8e00-f941b7850e12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.419639 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t7jk\" (UniqueName: \"kubernetes.io/projected/9de2ae30-b9b5-4d18-8e00-f941b7850e12-kube-api-access-9t7jk\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.419664 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de2ae30-b9b5-4d18-8e00-f941b7850e12-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.614470 4741 generic.go:334] "Generic (PLEG): container finished" podID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerID="34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f" exitCode=0 Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.614510 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbkp5" event={"ID":"9de2ae30-b9b5-4d18-8e00-f941b7850e12","Type":"ContainerDied","Data":"34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f"} Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.614543 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbkp5" event={"ID":"9de2ae30-b9b5-4d18-8e00-f941b7850e12","Type":"ContainerDied","Data":"9802efe303d6ae489942f711932e821cf03d2f54da86b8e31888b49f640540d2"} Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.614561 4741 scope.go:117] "RemoveContainer" containerID="34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.614706 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbkp5" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.632296 4741 scope.go:117] "RemoveContainer" containerID="8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.642972 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbkp5"] Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.653361 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kbkp5"] Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.664561 4741 scope.go:117] "RemoveContainer" containerID="28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.693009 4741 scope.go:117] "RemoveContainer" containerID="34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f" Sep 29 20:23:15 crc kubenswrapper[4741]: E0929 20:23:15.693478 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f\": container with ID starting with 34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f not found: ID does not exist" containerID="34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.693589 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f"} err="failed to get container status \"34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f\": rpc error: code = NotFound desc = could not find container \"34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f\": container with ID starting with 34b6b13d80a82effeb8e8a206024021e2036aa6b40f2a3231c68d5b1ce4ac93f not found: ID does not exist" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.693671 4741 scope.go:117] "RemoveContainer" containerID="8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5" Sep 29 20:23:15 crc kubenswrapper[4741]: E0929 20:23:15.694598 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5\": container with ID starting with 8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5 not found: ID does not exist" containerID="8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.694658 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5"} err="failed to get container status \"8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5\": rpc error: code = NotFound desc = could not find container \"8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5\": container with ID starting with 8260b12a169039aa134df07dd97b37ec9f1c16aad05956a681e334117914fce5 not found: ID does not exist" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.694694 4741 scope.go:117] "RemoveContainer" containerID="28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a" Sep 29 20:23:15 crc kubenswrapper[4741]: E0929 20:23:15.694971 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a\": container with ID starting with 28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a not found: ID does not exist" containerID="28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a" Sep 29 20:23:15 crc kubenswrapper[4741]: I0929 20:23:15.695074 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a"} err="failed to get container status \"28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a\": rpc error: code = NotFound desc = could not find container \"28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a\": container with ID starting with 28cddd8e7af92a1423a5186b3dcda4827f3d98cc498634cffcf418bbd83a522a not found: ID does not exist" Sep 29 20:23:17 crc kubenswrapper[4741]: I0929 20:23:17.096268 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" path="/var/lib/kubelet/pods/9de2ae30-b9b5-4d18-8e00-f941b7850e12/volumes" Sep 29 20:23:25 crc kubenswrapper[4741]: I0929 20:23:25.954972 4741 scope.go:117] "RemoveContainer" containerID="bc08b41c9620475f74ab11f5fe156357925b0bb34e7c96e5b73edf6a30202be6" Sep 29 20:23:31 crc kubenswrapper[4741]: I0929 20:23:31.738834 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:23:31 crc kubenswrapper[4741]: I0929 20:23:31.739598 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:24:01 crc kubenswrapper[4741]: I0929 20:24:01.739618 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:24:01 crc kubenswrapper[4741]: I0929 20:24:01.740304 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:24:31 crc kubenswrapper[4741]: I0929 20:24:31.738597 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:24:31 crc kubenswrapper[4741]: I0929 20:24:31.739204 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:24:31 crc kubenswrapper[4741]: I0929 20:24:31.739258 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:24:31 crc kubenswrapper[4741]: I0929 20:24:31.739961 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:24:31 crc kubenswrapper[4741]: I0929 20:24:31.740030 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" gracePeriod=600 Sep 29 20:24:31 crc kubenswrapper[4741]: E0929 20:24:31.867925 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:24:32 crc kubenswrapper[4741]: I0929 20:24:32.272105 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" exitCode=0 Sep 29 20:24:32 crc kubenswrapper[4741]: I0929 20:24:32.272181 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2"} Sep 29 20:24:32 crc kubenswrapper[4741]: I0929 20:24:32.272251 4741 scope.go:117] "RemoveContainer" containerID="ccef0d12b1eb34848ee629bb3ec047bd5ae677186ad2661355829fe5d3833a26" Sep 29 20:24:32 crc kubenswrapper[4741]: I0929 20:24:32.272838 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:24:32 crc kubenswrapper[4741]: E0929 20:24:32.273121 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:24:47 crc kubenswrapper[4741]: I0929 20:24:47.086758 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:24:47 crc kubenswrapper[4741]: E0929 20:24:47.087824 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:25:02 crc kubenswrapper[4741]: I0929 20:25:02.086675 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:25:02 crc kubenswrapper[4741]: E0929 20:25:02.089816 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:25:14 crc kubenswrapper[4741]: I0929 20:25:14.086584 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:25:14 crc kubenswrapper[4741]: E0929 20:25:14.087385 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:25:29 crc kubenswrapper[4741]: I0929 20:25:29.091855 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:25:29 crc kubenswrapper[4741]: E0929 20:25:29.092636 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.841635 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r6sbs"] Sep 29 20:25:34 crc kubenswrapper[4741]: E0929 20:25:34.842570 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="extract-utilities" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.842589 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="extract-utilities" Sep 29 20:25:34 crc kubenswrapper[4741]: E0929 20:25:34.842624 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="extract-content" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.842634 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="extract-content" Sep 29 20:25:34 crc kubenswrapper[4741]: E0929 20:25:34.842646 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="registry-server" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.842656 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="registry-server" Sep 29 20:25:34 crc kubenswrapper[4741]: E0929 20:25:34.842680 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="226db4b2-8bb4-44de-a918-4d933b4c31ef" containerName="storage" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.842688 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="226db4b2-8bb4-44de-a918-4d933b4c31ef" containerName="storage" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.842855 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="226db4b2-8bb4-44de-a918-4d933b4c31ef" containerName="storage" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.842873 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9de2ae30-b9b5-4d18-8e00-f941b7850e12" containerName="registry-server" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.844181 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.851948 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6sbs"] Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.952170 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-utilities\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.952234 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-catalog-content\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:34 crc kubenswrapper[4741]: I0929 20:25:34.952256 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdcqs\" (UniqueName: \"kubernetes.io/projected/77964efe-6870-4018-ad88-e93436ec9ebe-kube-api-access-xdcqs\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.053478 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-utilities\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.053854 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-catalog-content\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.053884 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdcqs\" (UniqueName: \"kubernetes.io/projected/77964efe-6870-4018-ad88-e93436ec9ebe-kube-api-access-xdcqs\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.054014 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-utilities\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.054216 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-catalog-content\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.077061 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdcqs\" (UniqueName: \"kubernetes.io/projected/77964efe-6870-4018-ad88-e93436ec9ebe-kube-api-access-xdcqs\") pod \"redhat-marketplace-r6sbs\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.163552 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.604249 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6sbs"] Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.810176 4741 generic.go:334] "Generic (PLEG): container finished" podID="77964efe-6870-4018-ad88-e93436ec9ebe" containerID="8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf" exitCode=0 Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.810218 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6sbs" event={"ID":"77964efe-6870-4018-ad88-e93436ec9ebe","Type":"ContainerDied","Data":"8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf"} Sep 29 20:25:35 crc kubenswrapper[4741]: I0929 20:25:35.810249 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6sbs" event={"ID":"77964efe-6870-4018-ad88-e93436ec9ebe","Type":"ContainerStarted","Data":"e0462c23e81570c22e52100754929ae717f92509f8165d65869e6db70272a863"} Sep 29 20:25:36 crc kubenswrapper[4741]: I0929 20:25:36.818212 4741 generic.go:334] "Generic (PLEG): container finished" podID="77964efe-6870-4018-ad88-e93436ec9ebe" containerID="5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021" exitCode=0 Sep 29 20:25:36 crc kubenswrapper[4741]: I0929 20:25:36.818297 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6sbs" event={"ID":"77964efe-6870-4018-ad88-e93436ec9ebe","Type":"ContainerDied","Data":"5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021"} Sep 29 20:25:37 crc kubenswrapper[4741]: I0929 20:25:37.827485 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6sbs" event={"ID":"77964efe-6870-4018-ad88-e93436ec9ebe","Type":"ContainerStarted","Data":"d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5"} Sep 29 20:25:37 crc kubenswrapper[4741]: I0929 20:25:37.845243 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r6sbs" podStartSLOduration=2.437879765 podStartE2EDuration="3.84522641s" podCreationTimestamp="2025-09-29 20:25:34 +0000 UTC" firstStartedPulling="2025-09-29 20:25:35.811447533 +0000 UTC m=+4577.459236865" lastFinishedPulling="2025-09-29 20:25:37.218794168 +0000 UTC m=+4578.866583510" observedRunningTime="2025-09-29 20:25:37.842218345 +0000 UTC m=+4579.490007677" watchObservedRunningTime="2025-09-29 20:25:37.84522641 +0000 UTC m=+4579.493015742" Sep 29 20:25:41 crc kubenswrapper[4741]: I0929 20:25:41.086716 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:25:41 crc kubenswrapper[4741]: E0929 20:25:41.089181 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:25:45 crc kubenswrapper[4741]: I0929 20:25:45.164171 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:45 crc kubenswrapper[4741]: I0929 20:25:45.164659 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:45 crc kubenswrapper[4741]: I0929 20:25:45.218997 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:45 crc kubenswrapper[4741]: I0929 20:25:45.940210 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:45 crc kubenswrapper[4741]: I0929 20:25:45.988525 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6sbs"] Sep 29 20:25:47 crc kubenswrapper[4741]: I0929 20:25:47.918286 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r6sbs" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="registry-server" containerID="cri-o://d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5" gracePeriod=2 Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.349560 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.540628 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-catalog-content\") pod \"77964efe-6870-4018-ad88-e93436ec9ebe\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.540769 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdcqs\" (UniqueName: \"kubernetes.io/projected/77964efe-6870-4018-ad88-e93436ec9ebe-kube-api-access-xdcqs\") pod \"77964efe-6870-4018-ad88-e93436ec9ebe\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.540812 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-utilities\") pod \"77964efe-6870-4018-ad88-e93436ec9ebe\" (UID: \"77964efe-6870-4018-ad88-e93436ec9ebe\") " Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.542011 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-utilities" (OuterVolumeSpecName: "utilities") pod "77964efe-6870-4018-ad88-e93436ec9ebe" (UID: "77964efe-6870-4018-ad88-e93436ec9ebe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.565630 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77964efe-6870-4018-ad88-e93436ec9ebe" (UID: "77964efe-6870-4018-ad88-e93436ec9ebe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.642407 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.642449 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77964efe-6870-4018-ad88-e93436ec9ebe-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.886470 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77964efe-6870-4018-ad88-e93436ec9ebe-kube-api-access-xdcqs" (OuterVolumeSpecName: "kube-api-access-xdcqs") pod "77964efe-6870-4018-ad88-e93436ec9ebe" (UID: "77964efe-6870-4018-ad88-e93436ec9ebe"). InnerVolumeSpecName "kube-api-access-xdcqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.930856 4741 generic.go:334] "Generic (PLEG): container finished" podID="77964efe-6870-4018-ad88-e93436ec9ebe" containerID="d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5" exitCode=0 Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.930899 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6sbs" event={"ID":"77964efe-6870-4018-ad88-e93436ec9ebe","Type":"ContainerDied","Data":"d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5"} Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.930928 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6sbs" event={"ID":"77964efe-6870-4018-ad88-e93436ec9ebe","Type":"ContainerDied","Data":"e0462c23e81570c22e52100754929ae717f92509f8165d65869e6db70272a863"} Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.930948 4741 scope.go:117] "RemoveContainer" containerID="d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.931052 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6sbs" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.946228 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdcqs\" (UniqueName: \"kubernetes.io/projected/77964efe-6870-4018-ad88-e93436ec9ebe-kube-api-access-xdcqs\") on node \"crc\" DevicePath \"\"" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.972767 4741 scope.go:117] "RemoveContainer" containerID="5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021" Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.991837 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6sbs"] Sep 29 20:25:48 crc kubenswrapper[4741]: I0929 20:25:48.998572 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6sbs"] Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.099798 4741 scope.go:117] "RemoveContainer" containerID="8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf" Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.100432 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" path="/var/lib/kubelet/pods/77964efe-6870-4018-ad88-e93436ec9ebe/volumes" Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.119774 4741 scope.go:117] "RemoveContainer" containerID="d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5" Sep 29 20:25:49 crc kubenswrapper[4741]: E0929 20:25:49.120632 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5\": container with ID starting with d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5 not found: ID does not exist" containerID="d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5" Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.120730 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5"} err="failed to get container status \"d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5\": rpc error: code = NotFound desc = could not find container \"d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5\": container with ID starting with d85ca31394e85ba893e18a43c7da0fa48712c1b5699b209e3b4ac83101818ea5 not found: ID does not exist" Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.120800 4741 scope.go:117] "RemoveContainer" containerID="5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021" Sep 29 20:25:49 crc kubenswrapper[4741]: E0929 20:25:49.121469 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021\": container with ID starting with 5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021 not found: ID does not exist" containerID="5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021" Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.121500 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021"} err="failed to get container status \"5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021\": rpc error: code = NotFound desc = could not find container \"5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021\": container with ID starting with 5b518892b5ffc3c41bb2dfbced0c0c3fccf5222f0d7a0f189af21efe7bce5021 not found: ID does not exist" Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.121521 4741 scope.go:117] "RemoveContainer" containerID="8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf" Sep 29 20:25:49 crc kubenswrapper[4741]: E0929 20:25:49.122000 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf\": container with ID starting with 8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf not found: ID does not exist" containerID="8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf" Sep 29 20:25:49 crc kubenswrapper[4741]: I0929 20:25:49.122041 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf"} err="failed to get container status \"8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf\": rpc error: code = NotFound desc = could not find container \"8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf\": container with ID starting with 8241a7de7c8d47373f420ebd0df1b8addab244b66e509099532803537bf5e7cf not found: ID does not exist" Sep 29 20:25:53 crc kubenswrapper[4741]: I0929 20:25:53.085847 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:25:53 crc kubenswrapper[4741]: E0929 20:25:53.088126 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:26:06 crc kubenswrapper[4741]: I0929 20:26:06.086285 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:26:06 crc kubenswrapper[4741]: E0929 20:26:06.087271 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:26:20 crc kubenswrapper[4741]: I0929 20:26:20.085717 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:26:20 crc kubenswrapper[4741]: E0929 20:26:20.086495 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.872752 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-6q2mz"] Sep 29 20:26:25 crc kubenswrapper[4741]: E0929 20:26:25.873618 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="registry-server" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.873633 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="registry-server" Sep 29 20:26:25 crc kubenswrapper[4741]: E0929 20:26:25.873657 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="extract-content" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.873662 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="extract-content" Sep 29 20:26:25 crc kubenswrapper[4741]: E0929 20:26:25.873677 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="extract-utilities" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.873685 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="extract-utilities" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.873813 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="77964efe-6870-4018-ad88-e93436ec9ebe" containerName="registry-server" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.874566 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.876666 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-sdpcb" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.876763 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.876914 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.877170 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.877345 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.882365 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-6q2mz"] Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.919119 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.919173 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-config\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:25 crc kubenswrapper[4741]: I0929 20:26:25.919258 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dhht\" (UniqueName: \"kubernetes.io/projected/ec28873d-b499-475f-b177-3d4445786108-kube-api-access-9dhht\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.020300 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dhht\" (UniqueName: \"kubernetes.io/projected/ec28873d-b499-475f-b177-3d4445786108-kube-api-access-9dhht\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.020672 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.020694 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-config\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.021779 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-config\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.021792 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.045134 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dhht\" (UniqueName: \"kubernetes.io/projected/ec28873d-b499-475f-b177-3d4445786108-kube-api-access-9dhht\") pod \"dnsmasq-dns-5d7b5456f5-6q2mz\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.080093 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-vqkkg"] Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.081202 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.103615 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-vqkkg"] Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.195482 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.233344 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.233411 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkvwm\" (UniqueName: \"kubernetes.io/projected/7d69ece8-bfe3-4a23-8e34-e80a201f2061-kube-api-access-lkvwm\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.233468 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-config\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.341251 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.341645 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkvwm\" (UniqueName: \"kubernetes.io/projected/7d69ece8-bfe3-4a23-8e34-e80a201f2061-kube-api-access-lkvwm\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.341720 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-config\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.342498 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-config\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.342987 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.371942 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkvwm\" (UniqueName: \"kubernetes.io/projected/7d69ece8-bfe3-4a23-8e34-e80a201f2061-kube-api-access-lkvwm\") pod \"dnsmasq-dns-98ddfc8f-vqkkg\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.417062 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.682343 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-6q2mz"] Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.834652 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-vqkkg"] Sep 29 20:26:26 crc kubenswrapper[4741]: W0929 20:26:26.844673 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d69ece8_bfe3_4a23_8e34_e80a201f2061.slice/crio-e5fe1e318413f5785f7a67c5f8d93cbc533f97e3342f3dd273a705fb5c3287f0 WatchSource:0}: Error finding container e5fe1e318413f5785f7a67c5f8d93cbc533f97e3342f3dd273a705fb5c3287f0: Status 404 returned error can't find the container with id e5fe1e318413f5785f7a67c5f8d93cbc533f97e3342f3dd273a705fb5c3287f0 Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.983932 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.985375 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.987432 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qr29j" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.987744 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.987990 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.988266 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 29 20:26:26 crc kubenswrapper[4741]: I0929 20:26:26.988514 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.005737 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050527 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050577 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050601 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050694 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050743 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050828 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050857 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050905 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.050954 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th5pm\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-kube-api-access-th5pm\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.151677 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.151763 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th5pm\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-kube-api-access-th5pm\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.151896 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.151939 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.151971 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.152012 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.152085 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.152170 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.152206 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.152295 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.153371 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.154068 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.154792 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.156971 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.157007 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a42e11aa99c12dd0dd5677559c9b04cb9a1bdd96f6f884392cb227f7ef4450bb/globalmount\"" pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.157733 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.158095 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.160377 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.187152 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th5pm\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-kube-api-access-th5pm\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.191981 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.263055 4741 generic.go:334] "Generic (PLEG): container finished" podID="ec28873d-b499-475f-b177-3d4445786108" containerID="1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553" exitCode=0 Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.263137 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" event={"ID":"ec28873d-b499-475f-b177-3d4445786108","Type":"ContainerDied","Data":"1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553"} Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.263170 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" event={"ID":"ec28873d-b499-475f-b177-3d4445786108","Type":"ContainerStarted","Data":"a3756d419eccd1275e9cd79b6c96676de3a184984e29bf0ede8c65eea77c9e0c"} Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.266482 4741 generic.go:334] "Generic (PLEG): container finished" podID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerID="fd726f34fd432e407eea2e3d738dff277b681f55b014014a1eb1c92b3ae03109" exitCode=0 Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.266540 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" event={"ID":"7d69ece8-bfe3-4a23-8e34-e80a201f2061","Type":"ContainerDied","Data":"fd726f34fd432e407eea2e3d738dff277b681f55b014014a1eb1c92b3ae03109"} Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.266579 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" event={"ID":"7d69ece8-bfe3-4a23-8e34-e80a201f2061","Type":"ContainerStarted","Data":"e5fe1e318413f5785f7a67c5f8d93cbc533f97e3342f3dd273a705fb5c3287f0"} Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.290168 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.291744 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.295728 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.295729 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.295840 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.295728 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.296031 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-npklm" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.297002 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.311416 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.458569 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.459676 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.460171 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq7d6\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-kube-api-access-fq7d6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.460261 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.460462 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.460694 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee5d9afc-ddfb-430a-8b41-109890ca764f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.460778 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.460905 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.460948 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee5d9afc-ddfb-430a-8b41-109890ca764f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.562822 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.562916 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.562958 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq7d6\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-kube-api-access-fq7d6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.562986 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.563008 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.563086 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee5d9afc-ddfb-430a-8b41-109890ca764f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.563564 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.563724 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.563813 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.564008 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.564047 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee5d9afc-ddfb-430a-8b41-109890ca764f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.565621 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.565847 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.567139 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.567174 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/60e10927b48fa6b5060e2bea37dd5a6e86cf748edc34351f02bb0425ec0b4853/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.574082 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee5d9afc-ddfb-430a-8b41-109890ca764f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.574280 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.574754 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee5d9afc-ddfb-430a-8b41-109890ca764f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.580447 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq7d6\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-kube-api-access-fq7d6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.598972 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.634502 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:26:27 crc kubenswrapper[4741]: W0929 20:26:27.850733 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e511a9d_7030_46e3_b6ec_2d77ce5beb8e.slice/crio-423a0005b244a7895e5c18477d660e837323fd99f0d0458d32194de9629826dd WatchSource:0}: Error finding container 423a0005b244a7895e5c18477d660e837323fd99f0d0458d32194de9629826dd: Status 404 returned error can't find the container with id 423a0005b244a7895e5c18477d660e837323fd99f0d0458d32194de9629826dd Sep 29 20:26:27 crc kubenswrapper[4741]: I0929 20:26:27.854023 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.182939 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.277346 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" event={"ID":"7d69ece8-bfe3-4a23-8e34-e80a201f2061","Type":"ContainerStarted","Data":"3134c8b043a46b4ac6990d6bda58a17eacd24fe28dd9adc048895accb2e732b7"} Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.277464 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.280159 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e","Type":"ContainerStarted","Data":"9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e"} Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.280205 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e","Type":"ContainerStarted","Data":"423a0005b244a7895e5c18477d660e837323fd99f0d0458d32194de9629826dd"} Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.283352 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" event={"ID":"ec28873d-b499-475f-b177-3d4445786108","Type":"ContainerStarted","Data":"67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52"} Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.283965 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.285853 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee5d9afc-ddfb-430a-8b41-109890ca764f","Type":"ContainerStarted","Data":"215ac874ad9e9ca48ebd989914d95e46ccec054d1eb429426a6615ac2fc909bf"} Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.302426 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" podStartSLOduration=2.302402571 podStartE2EDuration="2.302402571s" podCreationTimestamp="2025-09-29 20:26:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:26:28.298794847 +0000 UTC m=+4629.946584179" watchObservedRunningTime="2025-09-29 20:26:28.302402571 +0000 UTC m=+4629.950191913" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.324154 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" podStartSLOduration=3.324136336 podStartE2EDuration="3.324136336s" podCreationTimestamp="2025-09-29 20:26:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:26:28.319214151 +0000 UTC m=+4629.967003483" watchObservedRunningTime="2025-09-29 20:26:28.324136336 +0000 UTC m=+4629.971925668" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.393912 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.395443 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.398752 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.404111 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.404533 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5bh97" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.404643 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.406659 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.410965 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.418701 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.554519 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.555706 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.558527 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.558765 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-bgjs9" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.566877 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577210 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577257 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f79ca3c3-1af6-4776-bc53-fffef15a23d9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577286 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577312 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577337 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6kmb\" (UniqueName: \"kubernetes.io/projected/f79ca3c3-1af6-4776-bc53-fffef15a23d9-kube-api-access-s6kmb\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577359 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577400 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577425 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.577449 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678370 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6kmb\" (UniqueName: \"kubernetes.io/projected/f79ca3c3-1af6-4776-bc53-fffef15a23d9-kube-api-access-s6kmb\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678426 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678458 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678480 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678505 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e2f180-603f-4760-8506-eb10f26199fe-config-data\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678530 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678573 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678593 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f79ca3c3-1af6-4776-bc53-fffef15a23d9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678611 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678628 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7skb\" (UniqueName: \"kubernetes.io/projected/b1e2f180-603f-4760-8506-eb10f26199fe-kube-api-access-j7skb\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678651 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.678669 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1e2f180-603f-4760-8506-eb10f26199fe-kolla-config\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.679142 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f79ca3c3-1af6-4776-bc53-fffef15a23d9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.679833 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.679956 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.680442 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79ca3c3-1af6-4776-bc53-fffef15a23d9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.681953 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.682139 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.682167 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ee54bf1b298aaab4e92f648e76f16886299dfbbdd8f363a98152245d69d11ae8/globalmount\"" pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.682282 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.682838 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f79ca3c3-1af6-4776-bc53-fffef15a23d9-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.699659 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6kmb\" (UniqueName: \"kubernetes.io/projected/f79ca3c3-1af6-4776-bc53-fffef15a23d9-kube-api-access-s6kmb\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.712875 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6ee262f-a7e9-41a2-93fc-8bcb33332e97\") pod \"openstack-cell1-galera-0\" (UID: \"f79ca3c3-1af6-4776-bc53-fffef15a23d9\") " pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.721296 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.779762 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e2f180-603f-4760-8506-eb10f26199fe-config-data\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.779868 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7skb\" (UniqueName: \"kubernetes.io/projected/b1e2f180-603f-4760-8506-eb10f26199fe-kube-api-access-j7skb\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.779928 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1e2f180-603f-4760-8506-eb10f26199fe-kolla-config\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.781755 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e2f180-603f-4760-8506-eb10f26199fe-config-data\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.783712 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1e2f180-603f-4760-8506-eb10f26199fe-kolla-config\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.803099 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7skb\" (UniqueName: \"kubernetes.io/projected/b1e2f180-603f-4760-8506-eb10f26199fe-kube-api-access-j7skb\") pod \"memcached-0\" (UID: \"b1e2f180-603f-4760-8506-eb10f26199fe\") " pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.873494 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 20:26:28 crc kubenswrapper[4741]: I0929 20:26:28.959963 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.298787 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f79ca3c3-1af6-4776-bc53-fffef15a23d9","Type":"ContainerStarted","Data":"63d7f1f38f055d883d5f21b4305fbd6b65dc9abfa9895210a074cd457fabe3eb"} Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.298831 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f79ca3c3-1af6-4776-bc53-fffef15a23d9","Type":"ContainerStarted","Data":"2d67601508b2733534b788f42cb5a3bb390f9232b40b583d0a4f5747248b89c4"} Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.303835 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee5d9afc-ddfb-430a-8b41-109890ca764f","Type":"ContainerStarted","Data":"f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a"} Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.307783 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 29 20:26:29 crc kubenswrapper[4741]: W0929 20:26:29.312112 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1e2f180_603f_4760_8506_eb10f26199fe.slice/crio-de4ec6b854b713820a629e69b1b5f0e7954b3398529a44a3111054ee6a16f431 WatchSource:0}: Error finding container de4ec6b854b713820a629e69b1b5f0e7954b3398529a44a3111054ee6a16f431: Status 404 returned error can't find the container with id de4ec6b854b713820a629e69b1b5f0e7954b3398529a44a3111054ee6a16f431 Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.542052 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.546721 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.549076 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.549108 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.549618 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.549961 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6nj72" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.550089 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.696220 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.696285 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.696309 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b041ad-6a3d-43f2-8faa-392b7c05c539-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.696335 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.697895 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.697977 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-config-data-default\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.697997 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-secrets\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.698061 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5445\" (UniqueName: \"kubernetes.io/projected/d2b041ad-6a3d-43f2-8faa-392b7c05c539-kube-api-access-z5445\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.698149 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-kolla-config\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799376 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b041ad-6a3d-43f2-8faa-392b7c05c539-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799470 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799503 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799536 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-config-data-default\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799560 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-secrets\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799620 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5445\" (UniqueName: \"kubernetes.io/projected/d2b041ad-6a3d-43f2-8faa-392b7c05c539-kube-api-access-z5445\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799673 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-kolla-config\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799750 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.799798 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.800818 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b041ad-6a3d-43f2-8faa-392b7c05c539-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.801235 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-kolla-config\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.801303 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-config-data-default\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.802166 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b041ad-6a3d-43f2-8faa-392b7c05c539-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.803522 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.803548 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/47c482cac969aa418b138ba01ae21431cc319e3d3bd28fa97b9e7c0920e46866/globalmount\"" pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.805422 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-secrets\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.805862 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.810251 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b041ad-6a3d-43f2-8faa-392b7c05c539-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.817320 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5445\" (UniqueName: \"kubernetes.io/projected/d2b041ad-6a3d-43f2-8faa-392b7c05c539-kube-api-access-z5445\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.838890 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8cafe0fe-c591-4c05-9005-d96c6d6d1d3b\") pod \"openstack-galera-0\" (UID: \"d2b041ad-6a3d-43f2-8faa-392b7c05c539\") " pod="openstack/openstack-galera-0" Sep 29 20:26:29 crc kubenswrapper[4741]: I0929 20:26:29.862386 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 20:26:30 crc kubenswrapper[4741]: I0929 20:26:30.148990 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 20:26:30 crc kubenswrapper[4741]: I0929 20:26:30.310341 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d2b041ad-6a3d-43f2-8faa-392b7c05c539","Type":"ContainerStarted","Data":"9b05138eeae9ba5e83db2ddf0bc3d42073b6474c39bc81f1fe7d761b234a50b5"} Sep 29 20:26:30 crc kubenswrapper[4741]: I0929 20:26:30.312334 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b1e2f180-603f-4760-8506-eb10f26199fe","Type":"ContainerStarted","Data":"83c95f9b9fd11eff3ad0e12bbc6fcd4616b32df64a4a4dad7489805233a25f35"} Sep 29 20:26:30 crc kubenswrapper[4741]: I0929 20:26:30.312460 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b1e2f180-603f-4760-8506-eb10f26199fe","Type":"ContainerStarted","Data":"de4ec6b854b713820a629e69b1b5f0e7954b3398529a44a3111054ee6a16f431"} Sep 29 20:26:30 crc kubenswrapper[4741]: I0929 20:26:30.335586 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.335565567 podStartE2EDuration="2.335565567s" podCreationTimestamp="2025-09-29 20:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:26:30.331360385 +0000 UTC m=+4631.979149757" watchObservedRunningTime="2025-09-29 20:26:30.335565567 +0000 UTC m=+4631.983354909" Sep 29 20:26:31 crc kubenswrapper[4741]: I0929 20:26:31.086113 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:26:31 crc kubenswrapper[4741]: E0929 20:26:31.086370 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:26:31 crc kubenswrapper[4741]: I0929 20:26:31.322251 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d2b041ad-6a3d-43f2-8faa-392b7c05c539","Type":"ContainerStarted","Data":"9d9f381c7e9d31e1acb99672437f5e34e8589850fc70535c67845be887072c22"} Sep 29 20:26:31 crc kubenswrapper[4741]: I0929 20:26:31.323016 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Sep 29 20:26:33 crc kubenswrapper[4741]: I0929 20:26:33.340480 4741 generic.go:334] "Generic (PLEG): container finished" podID="f79ca3c3-1af6-4776-bc53-fffef15a23d9" containerID="63d7f1f38f055d883d5f21b4305fbd6b65dc9abfa9895210a074cd457fabe3eb" exitCode=0 Sep 29 20:26:33 crc kubenswrapper[4741]: I0929 20:26:33.340614 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f79ca3c3-1af6-4776-bc53-fffef15a23d9","Type":"ContainerDied","Data":"63d7f1f38f055d883d5f21b4305fbd6b65dc9abfa9895210a074cd457fabe3eb"} Sep 29 20:26:34 crc kubenswrapper[4741]: I0929 20:26:34.349914 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f79ca3c3-1af6-4776-bc53-fffef15a23d9","Type":"ContainerStarted","Data":"fcf85dfff3619f6aaa88e629cca018ee7282c2a4bffb9ca52b5f790c26418fce"} Sep 29 20:26:34 crc kubenswrapper[4741]: I0929 20:26:34.352033 4741 generic.go:334] "Generic (PLEG): container finished" podID="d2b041ad-6a3d-43f2-8faa-392b7c05c539" containerID="9d9f381c7e9d31e1acb99672437f5e34e8589850fc70535c67845be887072c22" exitCode=0 Sep 29 20:26:34 crc kubenswrapper[4741]: I0929 20:26:34.352076 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d2b041ad-6a3d-43f2-8faa-392b7c05c539","Type":"ContainerDied","Data":"9d9f381c7e9d31e1acb99672437f5e34e8589850fc70535c67845be887072c22"} Sep 29 20:26:34 crc kubenswrapper[4741]: I0929 20:26:34.378684 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.378665821 podStartE2EDuration="7.378665821s" podCreationTimestamp="2025-09-29 20:26:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:26:34.37450936 +0000 UTC m=+4636.022298702" watchObservedRunningTime="2025-09-29 20:26:34.378665821 +0000 UTC m=+4636.026455153" Sep 29 20:26:35 crc kubenswrapper[4741]: I0929 20:26:35.361847 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d2b041ad-6a3d-43f2-8faa-392b7c05c539","Type":"ContainerStarted","Data":"0d012400a4aa4776350328574cb224a6f06f10d5c3ffbc8fdcfbe58b95cdafc2"} Sep 29 20:26:35 crc kubenswrapper[4741]: I0929 20:26:35.382797 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.382777491 podStartE2EDuration="7.382777491s" podCreationTimestamp="2025-09-29 20:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:26:35.380069496 +0000 UTC m=+4637.027858838" watchObservedRunningTime="2025-09-29 20:26:35.382777491 +0000 UTC m=+4637.030566833" Sep 29 20:26:36 crc kubenswrapper[4741]: I0929 20:26:36.197770 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:36 crc kubenswrapper[4741]: I0929 20:26:36.419517 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:26:36 crc kubenswrapper[4741]: I0929 20:26:36.479655 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-6q2mz"] Sep 29 20:26:36 crc kubenswrapper[4741]: I0929 20:26:36.480258 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" podUID="ec28873d-b499-475f-b177-3d4445786108" containerName="dnsmasq-dns" containerID="cri-o://67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52" gracePeriod=10 Sep 29 20:26:36 crc kubenswrapper[4741]: I0929 20:26:36.891294 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.015565 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-dns-svc\") pod \"ec28873d-b499-475f-b177-3d4445786108\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.015620 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-config\") pod \"ec28873d-b499-475f-b177-3d4445786108\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.015656 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dhht\" (UniqueName: \"kubernetes.io/projected/ec28873d-b499-475f-b177-3d4445786108-kube-api-access-9dhht\") pod \"ec28873d-b499-475f-b177-3d4445786108\" (UID: \"ec28873d-b499-475f-b177-3d4445786108\") " Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.020484 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec28873d-b499-475f-b177-3d4445786108-kube-api-access-9dhht" (OuterVolumeSpecName: "kube-api-access-9dhht") pod "ec28873d-b499-475f-b177-3d4445786108" (UID: "ec28873d-b499-475f-b177-3d4445786108"). InnerVolumeSpecName "kube-api-access-9dhht". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.049527 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-config" (OuterVolumeSpecName: "config") pod "ec28873d-b499-475f-b177-3d4445786108" (UID: "ec28873d-b499-475f-b177-3d4445786108"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.055488 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec28873d-b499-475f-b177-3d4445786108" (UID: "ec28873d-b499-475f-b177-3d4445786108"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.117265 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.117657 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec28873d-b499-475f-b177-3d4445786108-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.117711 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dhht\" (UniqueName: \"kubernetes.io/projected/ec28873d-b499-475f-b177-3d4445786108-kube-api-access-9dhht\") on node \"crc\" DevicePath \"\"" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.378343 4741 generic.go:334] "Generic (PLEG): container finished" podID="ec28873d-b499-475f-b177-3d4445786108" containerID="67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52" exitCode=0 Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.378428 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" event={"ID":"ec28873d-b499-475f-b177-3d4445786108","Type":"ContainerDied","Data":"67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52"} Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.378535 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.379020 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-6q2mz" event={"ID":"ec28873d-b499-475f-b177-3d4445786108","Type":"ContainerDied","Data":"a3756d419eccd1275e9cd79b6c96676de3a184984e29bf0ede8c65eea77c9e0c"} Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.379070 4741 scope.go:117] "RemoveContainer" containerID="67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.398448 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-6q2mz"] Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.403131 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-6q2mz"] Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.404509 4741 scope.go:117] "RemoveContainer" containerID="1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.420849 4741 scope.go:117] "RemoveContainer" containerID="67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52" Sep 29 20:26:37 crc kubenswrapper[4741]: E0929 20:26:37.421259 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52\": container with ID starting with 67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52 not found: ID does not exist" containerID="67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.421321 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52"} err="failed to get container status \"67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52\": rpc error: code = NotFound desc = could not find container \"67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52\": container with ID starting with 67cc286d596343b570d6a45cfe459373308d22270691acacaee91afa09ca4e52 not found: ID does not exist" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.421358 4741 scope.go:117] "RemoveContainer" containerID="1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553" Sep 29 20:26:37 crc kubenswrapper[4741]: E0929 20:26:37.421719 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553\": container with ID starting with 1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553 not found: ID does not exist" containerID="1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553" Sep 29 20:26:37 crc kubenswrapper[4741]: I0929 20:26:37.421755 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553"} err="failed to get container status \"1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553\": rpc error: code = NotFound desc = could not find container \"1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553\": container with ID starting with 1ab89147d7fbeac1beaa3e39602351067125b24ea8c4402fd97c2f5f313ca553 not found: ID does not exist" Sep 29 20:26:38 crc kubenswrapper[4741]: I0929 20:26:38.721594 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:38 crc kubenswrapper[4741]: I0929 20:26:38.723631 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:38 crc kubenswrapper[4741]: I0929 20:26:38.874555 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Sep 29 20:26:39 crc kubenswrapper[4741]: I0929 20:26:39.095774 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec28873d-b499-475f-b177-3d4445786108" path="/var/lib/kubelet/pods/ec28873d-b499-475f-b177-3d4445786108/volumes" Sep 29 20:26:39 crc kubenswrapper[4741]: I0929 20:26:39.862628 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Sep 29 20:26:39 crc kubenswrapper[4741]: I0929 20:26:39.862674 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Sep 29 20:26:39 crc kubenswrapper[4741]: I0929 20:26:39.907783 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Sep 29 20:26:40 crc kubenswrapper[4741]: I0929 20:26:40.449901 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Sep 29 20:26:40 crc kubenswrapper[4741]: I0929 20:26:40.772751 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:40 crc kubenswrapper[4741]: I0929 20:26:40.814503 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Sep 29 20:26:45 crc kubenswrapper[4741]: I0929 20:26:45.086243 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:26:45 crc kubenswrapper[4741]: E0929 20:26:45.086838 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:26:59 crc kubenswrapper[4741]: I0929 20:26:59.094508 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:26:59 crc kubenswrapper[4741]: E0929 20:26:59.096575 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:26:59 crc kubenswrapper[4741]: I0929 20:26:59.562939 4741 generic.go:334] "Generic (PLEG): container finished" podID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerID="9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e" exitCode=0 Sep 29 20:26:59 crc kubenswrapper[4741]: I0929 20:26:59.563024 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e","Type":"ContainerDied","Data":"9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e"} Sep 29 20:26:59 crc kubenswrapper[4741]: I0929 20:26:59.569688 4741 generic.go:334] "Generic (PLEG): container finished" podID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerID="f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a" exitCode=0 Sep 29 20:26:59 crc kubenswrapper[4741]: I0929 20:26:59.569757 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee5d9afc-ddfb-430a-8b41-109890ca764f","Type":"ContainerDied","Data":"f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a"} Sep 29 20:27:00 crc kubenswrapper[4741]: I0929 20:27:00.579163 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee5d9afc-ddfb-430a-8b41-109890ca764f","Type":"ContainerStarted","Data":"1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91"} Sep 29 20:27:00 crc kubenswrapper[4741]: I0929 20:27:00.579618 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:00 crc kubenswrapper[4741]: I0929 20:27:00.580960 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e","Type":"ContainerStarted","Data":"f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c"} Sep 29 20:27:00 crc kubenswrapper[4741]: I0929 20:27:00.581101 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 29 20:27:00 crc kubenswrapper[4741]: I0929 20:27:00.599663 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=34.599645814 podStartE2EDuration="34.599645814s" podCreationTimestamp="2025-09-29 20:26:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:27:00.595335529 +0000 UTC m=+4662.243124861" watchObservedRunningTime="2025-09-29 20:27:00.599645814 +0000 UTC m=+4662.247435146" Sep 29 20:27:00 crc kubenswrapper[4741]: I0929 20:27:00.619570 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.619545922 podStartE2EDuration="35.619545922s" podCreationTimestamp="2025-09-29 20:26:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:27:00.61538243 +0000 UTC m=+4662.263171762" watchObservedRunningTime="2025-09-29 20:27:00.619545922 +0000 UTC m=+4662.267335284" Sep 29 20:27:11 crc kubenswrapper[4741]: I0929 20:27:11.086691 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:27:11 crc kubenswrapper[4741]: E0929 20:27:11.087818 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:27:17 crc kubenswrapper[4741]: I0929 20:27:17.314589 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 29 20:27:17 crc kubenswrapper[4741]: I0929 20:27:17.637577 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:23 crc kubenswrapper[4741]: I0929 20:27:23.086426 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:27:23 crc kubenswrapper[4741]: E0929 20:27:23.087527 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.200539 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-5dltm"] Sep 29 20:27:24 crc kubenswrapper[4741]: E0929 20:27:24.200909 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec28873d-b499-475f-b177-3d4445786108" containerName="init" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.200927 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec28873d-b499-475f-b177-3d4445786108" containerName="init" Sep 29 20:27:24 crc kubenswrapper[4741]: E0929 20:27:24.200946 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec28873d-b499-475f-b177-3d4445786108" containerName="dnsmasq-dns" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.200952 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec28873d-b499-475f-b177-3d4445786108" containerName="dnsmasq-dns" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.201142 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec28873d-b499-475f-b177-3d4445786108" containerName="dnsmasq-dns" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.202098 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.208137 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-5dltm"] Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.322174 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-config\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.322278 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4h9q\" (UniqueName: \"kubernetes.io/projected/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-kube-api-access-h4h9q\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.322319 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.424100 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4h9q\" (UniqueName: \"kubernetes.io/projected/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-kube-api-access-h4h9q\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.424165 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.424199 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-config\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.425015 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-config\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.425174 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.445695 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4h9q\" (UniqueName: \"kubernetes.io/projected/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-kube-api-access-h4h9q\") pod \"dnsmasq-dns-5b7946d7b9-5dltm\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.525081 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:24 crc kubenswrapper[4741]: I0929 20:27:24.983369 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-5dltm"] Sep 29 20:27:25 crc kubenswrapper[4741]: I0929 20:27:25.031324 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:27:25 crc kubenswrapper[4741]: I0929 20:27:25.666744 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:27:25 crc kubenswrapper[4741]: I0929 20:27:25.837645 4741 generic.go:334] "Generic (PLEG): container finished" podID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerID="02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6" exitCode=0 Sep 29 20:27:25 crc kubenswrapper[4741]: I0929 20:27:25.837749 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" event={"ID":"f2e5500e-0a6f-4ba9-a38e-7339cd30e215","Type":"ContainerDied","Data":"02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6"} Sep 29 20:27:25 crc kubenswrapper[4741]: I0929 20:27:25.837940 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" event={"ID":"f2e5500e-0a6f-4ba9-a38e-7339cd30e215","Type":"ContainerStarted","Data":"db50541e1de20661f73a1482dce47c6afe5385c9c1f6d4a4f32c3ede80a31071"} Sep 29 20:27:26 crc kubenswrapper[4741]: I0929 20:27:26.734220 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerName="rabbitmq" containerID="cri-o://f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c" gracePeriod=604799 Sep 29 20:27:26 crc kubenswrapper[4741]: I0929 20:27:26.846599 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" event={"ID":"f2e5500e-0a6f-4ba9-a38e-7339cd30e215","Type":"ContainerStarted","Data":"22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b"} Sep 29 20:27:26 crc kubenswrapper[4741]: I0929 20:27:26.846750 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:26 crc kubenswrapper[4741]: I0929 20:27:26.864742 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" podStartSLOduration=2.864724818 podStartE2EDuration="2.864724818s" podCreationTimestamp="2025-09-29 20:27:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:27:26.860138183 +0000 UTC m=+4688.507927526" watchObservedRunningTime="2025-09-29 20:27:26.864724818 +0000 UTC m=+4688.512514160" Sep 29 20:27:27 crc kubenswrapper[4741]: I0929 20:27:27.312427 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.234:5672: connect: connection refused" Sep 29 20:27:27 crc kubenswrapper[4741]: I0929 20:27:27.342989 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerName="rabbitmq" containerID="cri-o://1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91" gracePeriod=604799 Sep 29 20:27:27 crc kubenswrapper[4741]: I0929 20:27:27.635845 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.235:5672: connect: connection refused" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.275134 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356564 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-plugins-conf\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356643 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-erlang-cookie\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356778 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356816 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-plugins\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356859 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-server-conf\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356887 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-pod-info\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356927 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th5pm\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-kube-api-access-th5pm\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.356971 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-erlang-cookie-secret\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.357019 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-confd\") pod \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\" (UID: \"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.357266 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.357493 4741 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.357547 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.358258 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.363134 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.363653 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-kube-api-access-th5pm" (OuterVolumeSpecName: "kube-api-access-th5pm") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "kube-api-access-th5pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.373934 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-pod-info" (OuterVolumeSpecName: "pod-info") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.380349 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-server-conf" (OuterVolumeSpecName: "server-conf") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.385231 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f" (OuterVolumeSpecName: "persistence") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "pvc-b46505fc-a529-4faf-9354-44c850f4458f". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.459006 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") on node \"crc\" " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.459037 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.459050 4741 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-server-conf\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.459058 4741 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-pod-info\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.459066 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th5pm\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-kube-api-access-th5pm\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.459075 4741 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.459083 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.460739 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" (UID: "1e511a9d-7030-46e3-b6ec-2d77ce5beb8e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.489546 4741 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.489733 4741 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b46505fc-a529-4faf-9354-44c850f4458f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f") on node "crc" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.561024 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.561060 4741 reconciler_common.go:293] "Volume detached for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.827038 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.908668 4741 generic.go:334] "Generic (PLEG): container finished" podID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerID="f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c" exitCode=0 Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.908734 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e","Type":"ContainerDied","Data":"f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c"} Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.908763 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e511a9d-7030-46e3-b6ec-2d77ce5beb8e","Type":"ContainerDied","Data":"423a0005b244a7895e5c18477d660e837323fd99f0d0458d32194de9629826dd"} Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.908772 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.908781 4741 scope.go:117] "RemoveContainer" containerID="f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.912855 4741 generic.go:334] "Generic (PLEG): container finished" podID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerID="1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91" exitCode=0 Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.912896 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee5d9afc-ddfb-430a-8b41-109890ca764f","Type":"ContainerDied","Data":"1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91"} Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.912932 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.912943 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee5d9afc-ddfb-430a-8b41-109890ca764f","Type":"ContainerDied","Data":"215ac874ad9e9ca48ebd989914d95e46ccec054d1eb429426a6615ac2fc909bf"} Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.937244 4741 scope.go:117] "RemoveContainer" containerID="9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.964135 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.967752 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee5d9afc-ddfb-430a-8b41-109890ca764f-pod-info\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.967861 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-plugins\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.967904 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee5d9afc-ddfb-430a-8b41-109890ca764f-erlang-cookie-secret\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.967928 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-server-conf\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.968044 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.968073 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq7d6\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-kube-api-access-fq7d6\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.968121 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-plugins-conf\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.968136 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-confd\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.968167 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-erlang-cookie\") pod \"ee5d9afc-ddfb-430a-8b41-109890ca764f\" (UID: \"ee5d9afc-ddfb-430a-8b41-109890ca764f\") " Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.968803 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.969962 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.971609 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ee5d9afc-ddfb-430a-8b41-109890ca764f-pod-info" (OuterVolumeSpecName: "pod-info") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.975856 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.977241 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.977628 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-kube-api-access-fq7d6" (OuterVolumeSpecName: "kube-api-access-fq7d6") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "kube-api-access-fq7d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.981542 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee5d9afc-ddfb-430a-8b41-109890ca764f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.993790 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d" (OuterVolumeSpecName: "persistence") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.995041 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:27:33 crc kubenswrapper[4741]: E0929 20:27:33.995479 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerName="rabbitmq" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.995501 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerName="rabbitmq" Sep 29 20:27:33 crc kubenswrapper[4741]: E0929 20:27:33.995535 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerName="rabbitmq" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.995544 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerName="rabbitmq" Sep 29 20:27:33 crc kubenswrapper[4741]: E0929 20:27:33.995560 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerName="setup-container" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.995568 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerName="setup-container" Sep 29 20:27:33 crc kubenswrapper[4741]: E0929 20:27:33.995585 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerName="setup-container" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.995592 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerName="setup-container" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.995768 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" containerName="rabbitmq" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.995796 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" containerName="rabbitmq" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.997152 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.999014 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 29 20:27:33 crc kubenswrapper[4741]: I0929 20:27:33.999479 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.003864 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-server-conf" (OuterVolumeSpecName: "server-conf") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.019820 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qr29j" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.020016 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.020136 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.037182 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.069725 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.069766 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.069803 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.069829 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b838fcc-8842-45a4-bbfe-29db7765bda8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.069851 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b838fcc-8842-45a4-bbfe-29db7765bda8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070001 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070046 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b838fcc-8842-45a4-bbfe-29db7765bda8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070108 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b838fcc-8842-45a4-bbfe-29db7765bda8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070168 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65zp7\" (UniqueName: \"kubernetes.io/projected/7b838fcc-8842-45a4-bbfe-29db7765bda8-kube-api-access-65zp7\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070246 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070259 4741 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee5d9afc-ddfb-430a-8b41-109890ca764f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070270 4741 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-server-conf\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070291 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") on node \"crc\" " Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070301 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq7d6\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-kube-api-access-fq7d6\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070311 4741 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee5d9afc-ddfb-430a-8b41-109890ca764f-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070321 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.070330 4741 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee5d9afc-ddfb-430a-8b41-109890ca764f-pod-info\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.078626 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ee5d9afc-ddfb-430a-8b41-109890ca764f" (UID: "ee5d9afc-ddfb-430a-8b41-109890ca764f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.085515 4741 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.085725 4741 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d") on node "crc" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.110356 4741 scope.go:117] "RemoveContainer" containerID="f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c" Sep 29 20:27:34 crc kubenswrapper[4741]: E0929 20:27:34.110819 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c\": container with ID starting with f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c not found: ID does not exist" containerID="f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.110864 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c"} err="failed to get container status \"f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c\": rpc error: code = NotFound desc = could not find container \"f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c\": container with ID starting with f743b9b9f27f3b21563e050fd4454e625b9d3b17c0133b836c70064cfdc6504c not found: ID does not exist" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.110892 4741 scope.go:117] "RemoveContainer" containerID="9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e" Sep 29 20:27:34 crc kubenswrapper[4741]: E0929 20:27:34.111171 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e\": container with ID starting with 9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e not found: ID does not exist" containerID="9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.111208 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e"} err="failed to get container status \"9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e\": rpc error: code = NotFound desc = could not find container \"9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e\": container with ID starting with 9ac6f902a125c9810d347c1b74aef69f3b231fef62fbea5e6a17263acca7988e not found: ID does not exist" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.111234 4741 scope.go:117] "RemoveContainer" containerID="1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.125091 4741 scope.go:117] "RemoveContainer" containerID="f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.139616 4741 scope.go:117] "RemoveContainer" containerID="1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91" Sep 29 20:27:34 crc kubenswrapper[4741]: E0929 20:27:34.140045 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91\": container with ID starting with 1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91 not found: ID does not exist" containerID="1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.140098 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91"} err="failed to get container status \"1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91\": rpc error: code = NotFound desc = could not find container \"1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91\": container with ID starting with 1da7fc2d8e8fa52e96799b29895455194c0abda8791112d1191e22e03fbdab91 not found: ID does not exist" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.140131 4741 scope.go:117] "RemoveContainer" containerID="f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a" Sep 29 20:27:34 crc kubenswrapper[4741]: E0929 20:27:34.140449 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a\": container with ID starting with f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a not found: ID does not exist" containerID="f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.140475 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a"} err="failed to get container status \"f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a\": rpc error: code = NotFound desc = could not find container \"f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a\": container with ID starting with f4c4163268526ed88836ac4ac78a033defcbe44ad6a0bb0477f6abb46e0b142a not found: ID does not exist" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171530 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171567 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b838fcc-8842-45a4-bbfe-29db7765bda8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171606 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b838fcc-8842-45a4-bbfe-29db7765bda8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171651 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65zp7\" (UniqueName: \"kubernetes.io/projected/7b838fcc-8842-45a4-bbfe-29db7765bda8-kube-api-access-65zp7\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171686 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171709 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171751 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171779 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b838fcc-8842-45a4-bbfe-29db7765bda8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171810 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b838fcc-8842-45a4-bbfe-29db7765bda8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171870 4741 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee5d9afc-ddfb-430a-8b41-109890ca764f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.171889 4741 reconciler_common.go:293] "Volume detached for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.172350 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.172636 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.172846 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b838fcc-8842-45a4-bbfe-29db7765bda8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.172865 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b838fcc-8842-45a4-bbfe-29db7765bda8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.174858 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.175138 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a42e11aa99c12dd0dd5677559c9b04cb9a1bdd96f6f884392cb227f7ef4450bb/globalmount\"" pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.175546 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b838fcc-8842-45a4-bbfe-29db7765bda8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.175602 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b838fcc-8842-45a4-bbfe-29db7765bda8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.177056 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b838fcc-8842-45a4-bbfe-29db7765bda8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.187722 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65zp7\" (UniqueName: \"kubernetes.io/projected/7b838fcc-8842-45a4-bbfe-29db7765bda8-kube-api-access-65zp7\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.201477 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b46505fc-a529-4faf-9354-44c850f4458f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b46505fc-a529-4faf-9354-44c850f4458f\") pod \"rabbitmq-server-0\" (UID: \"7b838fcc-8842-45a4-bbfe-29db7765bda8\") " pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.243670 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.248920 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.262455 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.265896 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.268349 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.268758 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.269091 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.270251 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-npklm" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.270981 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.280555 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374590 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374634 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374678 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374709 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374727 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374846 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz7hk\" (UniqueName: \"kubernetes.io/projected/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-kube-api-access-jz7hk\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374880 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374908 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.374939 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.412471 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476376 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476429 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476456 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476484 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476505 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476542 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476566 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476581 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.476629 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz7hk\" (UniqueName: \"kubernetes.io/projected/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-kube-api-access-jz7hk\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.477168 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.477734 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.478051 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.478125 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.479583 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.479604 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/60e10927b48fa6b5060e2bea37dd5a6e86cf748edc34351f02bb0425ec0b4853/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.479895 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.480217 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.480886 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.498549 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz7hk\" (UniqueName: \"kubernetes.io/projected/b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844-kube-api-access-jz7hk\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.521195 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6534bbe-9656-4c85-a53c-ed8d6759393d\") pod \"rabbitmq-cell1-server-0\" (UID: \"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.526534 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.584007 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-vqkkg"] Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.584264 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" podUID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerName="dnsmasq-dns" containerID="cri-o://3134c8b043a46b4ac6990d6bda58a17eacd24fe28dd9adc048895accb2e732b7" gracePeriod=10 Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.590822 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.843210 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.929743 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b838fcc-8842-45a4-bbfe-29db7765bda8","Type":"ContainerStarted","Data":"a7d568ca4e22e3826c0484546241b684c0bb5e78cbeb8ed6b242f636a5cd1235"} Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.933375 4741 generic.go:334] "Generic (PLEG): container finished" podID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerID="3134c8b043a46b4ac6990d6bda58a17eacd24fe28dd9adc048895accb2e732b7" exitCode=0 Sep 29 20:27:34 crc kubenswrapper[4741]: I0929 20:27:34.933429 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" event={"ID":"7d69ece8-bfe3-4a23-8e34-e80a201f2061","Type":"ContainerDied","Data":"3134c8b043a46b4ac6990d6bda58a17eacd24fe28dd9adc048895accb2e732b7"} Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.040449 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.085163 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-config\") pod \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.085322 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkvwm\" (UniqueName: \"kubernetes.io/projected/7d69ece8-bfe3-4a23-8e34-e80a201f2061-kube-api-access-lkvwm\") pod \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.085370 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-dns-svc\") pod \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\" (UID: \"7d69ece8-bfe3-4a23-8e34-e80a201f2061\") " Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.090747 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d69ece8-bfe3-4a23-8e34-e80a201f2061-kube-api-access-lkvwm" (OuterVolumeSpecName: "kube-api-access-lkvwm") pod "7d69ece8-bfe3-4a23-8e34-e80a201f2061" (UID: "7d69ece8-bfe3-4a23-8e34-e80a201f2061"). InnerVolumeSpecName "kube-api-access-lkvwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.100983 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e511a9d-7030-46e3-b6ec-2d77ce5beb8e" path="/var/lib/kubelet/pods/1e511a9d-7030-46e3-b6ec-2d77ce5beb8e/volumes" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.102550 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee5d9afc-ddfb-430a-8b41-109890ca764f" path="/var/lib/kubelet/pods/ee5d9afc-ddfb-430a-8b41-109890ca764f/volumes" Sep 29 20:27:35 crc kubenswrapper[4741]: W0929 20:27:35.104076 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5e09dd9_bbbc_4efa_a28c_2d6aaf4fe844.slice/crio-dd9e88fff3fd89f79db6be2798242dc0aaad1cb38c26bcdb051d6647391d37a0 WatchSource:0}: Error finding container dd9e88fff3fd89f79db6be2798242dc0aaad1cb38c26bcdb051d6647391d37a0: Status 404 returned error can't find the container with id dd9e88fff3fd89f79db6be2798242dc0aaad1cb38c26bcdb051d6647391d37a0 Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.136326 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d69ece8-bfe3-4a23-8e34-e80a201f2061" (UID: "7d69ece8-bfe3-4a23-8e34-e80a201f2061"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.138635 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-config" (OuterVolumeSpecName: "config") pod "7d69ece8-bfe3-4a23-8e34-e80a201f2061" (UID: "7d69ece8-bfe3-4a23-8e34-e80a201f2061"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.187387 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkvwm\" (UniqueName: \"kubernetes.io/projected/7d69ece8-bfe3-4a23-8e34-e80a201f2061-kube-api-access-lkvwm\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.187434 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.187445 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d69ece8-bfe3-4a23-8e34-e80a201f2061-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.188004 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.944651 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844","Type":"ContainerStarted","Data":"88f8828c3e255b5a18654ea1298d55b789d5ee0c4cde183814038141feb00de6"} Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.944702 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844","Type":"ContainerStarted","Data":"dd9e88fff3fd89f79db6be2798242dc0aaad1cb38c26bcdb051d6647391d37a0"} Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.949493 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b838fcc-8842-45a4-bbfe-29db7765bda8","Type":"ContainerStarted","Data":"6d5f7c23668ee307b395f7e33bae1d5f2a144eb5b0d299db26a69d8db7c7daf2"} Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.952968 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" event={"ID":"7d69ece8-bfe3-4a23-8e34-e80a201f2061","Type":"ContainerDied","Data":"e5fe1e318413f5785f7a67c5f8d93cbc533f97e3342f3dd273a705fb5c3287f0"} Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.953038 4741 scope.go:117] "RemoveContainer" containerID="3134c8b043a46b4ac6990d6bda58a17eacd24fe28dd9adc048895accb2e732b7" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.953107 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-vqkkg" Sep 29 20:27:35 crc kubenswrapper[4741]: I0929 20:27:35.984467 4741 scope.go:117] "RemoveContainer" containerID="fd726f34fd432e407eea2e3d738dff277b681f55b014014a1eb1c92b3ae03109" Sep 29 20:27:36 crc kubenswrapper[4741]: I0929 20:27:36.056222 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-vqkkg"] Sep 29 20:27:36 crc kubenswrapper[4741]: I0929 20:27:36.064652 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-vqkkg"] Sep 29 20:27:37 crc kubenswrapper[4741]: I0929 20:27:37.088013 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:27:37 crc kubenswrapper[4741]: E0929 20:27:37.088381 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:27:37 crc kubenswrapper[4741]: I0929 20:27:37.098940 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" path="/var/lib/kubelet/pods/7d69ece8-bfe3-4a23-8e34-e80a201f2061/volumes" Sep 29 20:27:51 crc kubenswrapper[4741]: I0929 20:27:51.086782 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:27:51 crc kubenswrapper[4741]: E0929 20:27:51.088869 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:28:02 crc kubenswrapper[4741]: I0929 20:28:02.086223 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:28:02 crc kubenswrapper[4741]: E0929 20:28:02.087006 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:28:06 crc kubenswrapper[4741]: I0929 20:28:06.216115 4741 generic.go:334] "Generic (PLEG): container finished" podID="b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844" containerID="88f8828c3e255b5a18654ea1298d55b789d5ee0c4cde183814038141feb00de6" exitCode=0 Sep 29 20:28:06 crc kubenswrapper[4741]: I0929 20:28:06.216181 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844","Type":"ContainerDied","Data":"88f8828c3e255b5a18654ea1298d55b789d5ee0c4cde183814038141feb00de6"} Sep 29 20:28:06 crc kubenswrapper[4741]: I0929 20:28:06.219432 4741 generic.go:334] "Generic (PLEG): container finished" podID="7b838fcc-8842-45a4-bbfe-29db7765bda8" containerID="6d5f7c23668ee307b395f7e33bae1d5f2a144eb5b0d299db26a69d8db7c7daf2" exitCode=0 Sep 29 20:28:06 crc kubenswrapper[4741]: I0929 20:28:06.219487 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b838fcc-8842-45a4-bbfe-29db7765bda8","Type":"ContainerDied","Data":"6d5f7c23668ee307b395f7e33bae1d5f2a144eb5b0d299db26a69d8db7c7daf2"} Sep 29 20:28:07 crc kubenswrapper[4741]: I0929 20:28:07.227774 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844","Type":"ContainerStarted","Data":"449b2d1783e088f691a2f76d1b2e9910b96abf803b3e755758876cd1bc77b49d"} Sep 29 20:28:07 crc kubenswrapper[4741]: I0929 20:28:07.229205 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:28:07 crc kubenswrapper[4741]: I0929 20:28:07.230594 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b838fcc-8842-45a4-bbfe-29db7765bda8","Type":"ContainerStarted","Data":"140b576b6a5472395ad46f5f12853c7f2a0f50fc34ce930d8e66130c7ad2ae4e"} Sep 29 20:28:07 crc kubenswrapper[4741]: I0929 20:28:07.231217 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 29 20:28:07 crc kubenswrapper[4741]: I0929 20:28:07.268653 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=33.268623801 podStartE2EDuration="33.268623801s" podCreationTimestamp="2025-09-29 20:27:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:28:07.264313966 +0000 UTC m=+4728.912103318" watchObservedRunningTime="2025-09-29 20:28:07.268623801 +0000 UTC m=+4728.916413133" Sep 29 20:28:07 crc kubenswrapper[4741]: I0929 20:28:07.289153 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=34.289132278 podStartE2EDuration="34.289132278s" podCreationTimestamp="2025-09-29 20:27:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:28:07.285503523 +0000 UTC m=+4728.933292855" watchObservedRunningTime="2025-09-29 20:28:07.289132278 +0000 UTC m=+4728.936921610" Sep 29 20:28:15 crc kubenswrapper[4741]: I0929 20:28:15.086452 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:28:15 crc kubenswrapper[4741]: E0929 20:28:15.088092 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:28:24 crc kubenswrapper[4741]: I0929 20:28:24.415420 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 29 20:28:24 crc kubenswrapper[4741]: I0929 20:28:24.593564 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 29 20:28:28 crc kubenswrapper[4741]: I0929 20:28:28.085421 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:28:28 crc kubenswrapper[4741]: E0929 20:28:28.086577 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.758312 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Sep 29 20:28:32 crc kubenswrapper[4741]: E0929 20:28:32.759255 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerName="init" Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.759281 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerName="init" Sep 29 20:28:32 crc kubenswrapper[4741]: E0929 20:28:32.759311 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerName="dnsmasq-dns" Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.759323 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerName="dnsmasq-dns" Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.759702 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d69ece8-bfe3-4a23-8e34-e80a201f2061" containerName="dnsmasq-dns" Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.760636 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.766688 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t6ql4" Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.768854 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Sep 29 20:28:32 crc kubenswrapper[4741]: I0929 20:28:32.899820 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frpv7\" (UniqueName: \"kubernetes.io/projected/affa1c03-2669-4ba8-bfc4-92f1c5c76c5c-kube-api-access-frpv7\") pod \"mariadb-client-1-default\" (UID: \"affa1c03-2669-4ba8-bfc4-92f1c5c76c5c\") " pod="openstack/mariadb-client-1-default" Sep 29 20:28:33 crc kubenswrapper[4741]: I0929 20:28:33.001813 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frpv7\" (UniqueName: \"kubernetes.io/projected/affa1c03-2669-4ba8-bfc4-92f1c5c76c5c-kube-api-access-frpv7\") pod \"mariadb-client-1-default\" (UID: \"affa1c03-2669-4ba8-bfc4-92f1c5c76c5c\") " pod="openstack/mariadb-client-1-default" Sep 29 20:28:33 crc kubenswrapper[4741]: I0929 20:28:33.037978 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frpv7\" (UniqueName: \"kubernetes.io/projected/affa1c03-2669-4ba8-bfc4-92f1c5c76c5c-kube-api-access-frpv7\") pod \"mariadb-client-1-default\" (UID: \"affa1c03-2669-4ba8-bfc4-92f1c5c76c5c\") " pod="openstack/mariadb-client-1-default" Sep 29 20:28:33 crc kubenswrapper[4741]: I0929 20:28:33.097208 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Sep 29 20:28:33 crc kubenswrapper[4741]: I0929 20:28:33.722748 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Sep 29 20:28:34 crc kubenswrapper[4741]: I0929 20:28:34.456638 4741 generic.go:334] "Generic (PLEG): container finished" podID="affa1c03-2669-4ba8-bfc4-92f1c5c76c5c" containerID="2856bb2c0447bc7c79e50745235a9a547ffd3f7bc9456d3ae05bd43e3dd3b98c" exitCode=0 Sep 29 20:28:34 crc kubenswrapper[4741]: I0929 20:28:34.456719 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"affa1c03-2669-4ba8-bfc4-92f1c5c76c5c","Type":"ContainerDied","Data":"2856bb2c0447bc7c79e50745235a9a547ffd3f7bc9456d3ae05bd43e3dd3b98c"} Sep 29 20:28:34 crc kubenswrapper[4741]: I0929 20:28:34.458603 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"affa1c03-2669-4ba8-bfc4-92f1c5c76c5c","Type":"ContainerStarted","Data":"ed0bf9e5de851fbc93fe468da2976adb9a9f0981ab0df1e0151eec9c270841c6"} Sep 29 20:28:35 crc kubenswrapper[4741]: I0929 20:28:35.795118 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Sep 29 20:28:35 crc kubenswrapper[4741]: I0929 20:28:35.825111 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_affa1c03-2669-4ba8-bfc4-92f1c5c76c5c/mariadb-client-1-default/0.log" Sep 29 20:28:35 crc kubenswrapper[4741]: I0929 20:28:35.852197 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Sep 29 20:28:35 crc kubenswrapper[4741]: I0929 20:28:35.858587 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Sep 29 20:28:35 crc kubenswrapper[4741]: I0929 20:28:35.945913 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frpv7\" (UniqueName: \"kubernetes.io/projected/affa1c03-2669-4ba8-bfc4-92f1c5c76c5c-kube-api-access-frpv7\") pod \"affa1c03-2669-4ba8-bfc4-92f1c5c76c5c\" (UID: \"affa1c03-2669-4ba8-bfc4-92f1c5c76c5c\") " Sep 29 20:28:35 crc kubenswrapper[4741]: I0929 20:28:35.959253 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/affa1c03-2669-4ba8-bfc4-92f1c5c76c5c-kube-api-access-frpv7" (OuterVolumeSpecName: "kube-api-access-frpv7") pod "affa1c03-2669-4ba8-bfc4-92f1c5c76c5c" (UID: "affa1c03-2669-4ba8-bfc4-92f1c5c76c5c"). InnerVolumeSpecName "kube-api-access-frpv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.047779 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frpv7\" (UniqueName: \"kubernetes.io/projected/affa1c03-2669-4ba8-bfc4-92f1c5c76c5c-kube-api-access-frpv7\") on node \"crc\" DevicePath \"\"" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.453420 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Sep 29 20:28:36 crc kubenswrapper[4741]: E0929 20:28:36.453906 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affa1c03-2669-4ba8-bfc4-92f1c5c76c5c" containerName="mariadb-client-1-default" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.453934 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="affa1c03-2669-4ba8-bfc4-92f1c5c76c5c" containerName="mariadb-client-1-default" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.454366 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="affa1c03-2669-4ba8-bfc4-92f1c5c76c5c" containerName="mariadb-client-1-default" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.455277 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.464279 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.480947 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed0bf9e5de851fbc93fe468da2976adb9a9f0981ab0df1e0151eec9c270841c6" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.480990 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.556005 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j48ww\" (UniqueName: \"kubernetes.io/projected/44ec763c-e220-4a30-a1af-50c1335ca92d-kube-api-access-j48ww\") pod \"mariadb-client-2-default\" (UID: \"44ec763c-e220-4a30-a1af-50c1335ca92d\") " pod="openstack/mariadb-client-2-default" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.657457 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j48ww\" (UniqueName: \"kubernetes.io/projected/44ec763c-e220-4a30-a1af-50c1335ca92d-kube-api-access-j48ww\") pod \"mariadb-client-2-default\" (UID: \"44ec763c-e220-4a30-a1af-50c1335ca92d\") " pod="openstack/mariadb-client-2-default" Sep 29 20:28:36 crc kubenswrapper[4741]: I0929 20:28:36.887840 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j48ww\" (UniqueName: \"kubernetes.io/projected/44ec763c-e220-4a30-a1af-50c1335ca92d-kube-api-access-j48ww\") pod \"mariadb-client-2-default\" (UID: \"44ec763c-e220-4a30-a1af-50c1335ca92d\") " pod="openstack/mariadb-client-2-default" Sep 29 20:28:37 crc kubenswrapper[4741]: I0929 20:28:37.085824 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Sep 29 20:28:37 crc kubenswrapper[4741]: I0929 20:28:37.108328 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="affa1c03-2669-4ba8-bfc4-92f1c5c76c5c" path="/var/lib/kubelet/pods/affa1c03-2669-4ba8-bfc4-92f1c5c76c5c/volumes" Sep 29 20:28:37 crc kubenswrapper[4741]: I0929 20:28:37.688743 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Sep 29 20:28:37 crc kubenswrapper[4741]: W0929 20:28:37.692857 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44ec763c_e220_4a30_a1af_50c1335ca92d.slice/crio-7cf033e88c0648a06467c55638241b225edffcce01b6c2b73b08362e98c7121e WatchSource:0}: Error finding container 7cf033e88c0648a06467c55638241b225edffcce01b6c2b73b08362e98c7121e: Status 404 returned error can't find the container with id 7cf033e88c0648a06467c55638241b225edffcce01b6c2b73b08362e98c7121e Sep 29 20:28:38 crc kubenswrapper[4741]: I0929 20:28:38.502979 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"44ec763c-e220-4a30-a1af-50c1335ca92d","Type":"ContainerStarted","Data":"8a45d581e930c393e9fd1054830ac3b6bc91c0380424f51e48e73152d3ce6bf8"} Sep 29 20:28:38 crc kubenswrapper[4741]: I0929 20:28:38.503571 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"44ec763c-e220-4a30-a1af-50c1335ca92d","Type":"ContainerStarted","Data":"7cf033e88c0648a06467c55638241b225edffcce01b6c2b73b08362e98c7121e"} Sep 29 20:28:38 crc kubenswrapper[4741]: I0929 20:28:38.532222 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=2.532194049 podStartE2EDuration="2.532194049s" podCreationTimestamp="2025-09-29 20:28:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:28:38.523600248 +0000 UTC m=+4760.171389620" watchObservedRunningTime="2025-09-29 20:28:38.532194049 +0000 UTC m=+4760.179983421" Sep 29 20:28:39 crc kubenswrapper[4741]: I0929 20:28:39.514115 4741 generic.go:334] "Generic (PLEG): container finished" podID="44ec763c-e220-4a30-a1af-50c1335ca92d" containerID="8a45d581e930c393e9fd1054830ac3b6bc91c0380424f51e48e73152d3ce6bf8" exitCode=0 Sep 29 20:28:39 crc kubenswrapper[4741]: I0929 20:28:39.514192 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"44ec763c-e220-4a30-a1af-50c1335ca92d","Type":"ContainerDied","Data":"8a45d581e930c393e9fd1054830ac3b6bc91c0380424f51e48e73152d3ce6bf8"} Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.026862 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.069421 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.075317 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.086503 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:28:41 crc kubenswrapper[4741]: E0929 20:28:41.086863 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.129270 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j48ww\" (UniqueName: \"kubernetes.io/projected/44ec763c-e220-4a30-a1af-50c1335ca92d-kube-api-access-j48ww\") pod \"44ec763c-e220-4a30-a1af-50c1335ca92d\" (UID: \"44ec763c-e220-4a30-a1af-50c1335ca92d\") " Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.136557 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44ec763c-e220-4a30-a1af-50c1335ca92d-kube-api-access-j48ww" (OuterVolumeSpecName: "kube-api-access-j48ww") pod "44ec763c-e220-4a30-a1af-50c1335ca92d" (UID: "44ec763c-e220-4a30-a1af-50c1335ca92d"). InnerVolumeSpecName "kube-api-access-j48ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.231107 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j48ww\" (UniqueName: \"kubernetes.io/projected/44ec763c-e220-4a30-a1af-50c1335ca92d-kube-api-access-j48ww\") on node \"crc\" DevicePath \"\"" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.536758 4741 scope.go:117] "RemoveContainer" containerID="8a45d581e930c393e9fd1054830ac3b6bc91c0380424f51e48e73152d3ce6bf8" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.536861 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.685585 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Sep 29 20:28:41 crc kubenswrapper[4741]: E0929 20:28:41.686167 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ec763c-e220-4a30-a1af-50c1335ca92d" containerName="mariadb-client-2-default" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.686195 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ec763c-e220-4a30-a1af-50c1335ca92d" containerName="mariadb-client-2-default" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.686426 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="44ec763c-e220-4a30-a1af-50c1335ca92d" containerName="mariadb-client-2-default" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.688216 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.690625 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t6ql4" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.691485 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.737870 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqlqv\" (UniqueName: \"kubernetes.io/projected/44cfe60a-9917-4967-a852-73452c90a012-kube-api-access-nqlqv\") pod \"mariadb-client-1\" (UID: \"44cfe60a-9917-4967-a852-73452c90a012\") " pod="openstack/mariadb-client-1" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.839602 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqlqv\" (UniqueName: \"kubernetes.io/projected/44cfe60a-9917-4967-a852-73452c90a012-kube-api-access-nqlqv\") pod \"mariadb-client-1\" (UID: \"44cfe60a-9917-4967-a852-73452c90a012\") " pod="openstack/mariadb-client-1" Sep 29 20:28:41 crc kubenswrapper[4741]: I0929 20:28:41.856227 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqlqv\" (UniqueName: \"kubernetes.io/projected/44cfe60a-9917-4967-a852-73452c90a012-kube-api-access-nqlqv\") pod \"mariadb-client-1\" (UID: \"44cfe60a-9917-4967-a852-73452c90a012\") " pod="openstack/mariadb-client-1" Sep 29 20:28:42 crc kubenswrapper[4741]: I0929 20:28:42.022815 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Sep 29 20:28:42 crc kubenswrapper[4741]: I0929 20:28:42.557836 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Sep 29 20:28:42 crc kubenswrapper[4741]: W0929 20:28:42.567865 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44cfe60a_9917_4967_a852_73452c90a012.slice/crio-c8e1ab219ed0e9748c38ce85fd5b66919ef326c5b658e9b9df973c2f8fba31e4 WatchSource:0}: Error finding container c8e1ab219ed0e9748c38ce85fd5b66919ef326c5b658e9b9df973c2f8fba31e4: Status 404 returned error can't find the container with id c8e1ab219ed0e9748c38ce85fd5b66919ef326c5b658e9b9df973c2f8fba31e4 Sep 29 20:28:43 crc kubenswrapper[4741]: I0929 20:28:43.095499 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44ec763c-e220-4a30-a1af-50c1335ca92d" path="/var/lib/kubelet/pods/44ec763c-e220-4a30-a1af-50c1335ca92d/volumes" Sep 29 20:28:43 crc kubenswrapper[4741]: I0929 20:28:43.557756 4741 generic.go:334] "Generic (PLEG): container finished" podID="44cfe60a-9917-4967-a852-73452c90a012" containerID="521442e9a2af2eeadf5adde3e17ac3c09cd84ccb90d8dc8d9e7ab16e3bd6853f" exitCode=0 Sep 29 20:28:43 crc kubenswrapper[4741]: I0929 20:28:43.557802 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"44cfe60a-9917-4967-a852-73452c90a012","Type":"ContainerDied","Data":"521442e9a2af2eeadf5adde3e17ac3c09cd84ccb90d8dc8d9e7ab16e3bd6853f"} Sep 29 20:28:43 crc kubenswrapper[4741]: I0929 20:28:43.557833 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"44cfe60a-9917-4967-a852-73452c90a012","Type":"ContainerStarted","Data":"c8e1ab219ed0e9748c38ce85fd5b66919ef326c5b658e9b9df973c2f8fba31e4"} Sep 29 20:28:44 crc kubenswrapper[4741]: I0929 20:28:44.981104 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.000770 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_44cfe60a-9917-4967-a852-73452c90a012/mariadb-client-1/0.log" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.031102 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.038875 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.091786 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqlqv\" (UniqueName: \"kubernetes.io/projected/44cfe60a-9917-4967-a852-73452c90a012-kube-api-access-nqlqv\") pod \"44cfe60a-9917-4967-a852-73452c90a012\" (UID: \"44cfe60a-9917-4967-a852-73452c90a012\") " Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.098650 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44cfe60a-9917-4967-a852-73452c90a012-kube-api-access-nqlqv" (OuterVolumeSpecName: "kube-api-access-nqlqv") pod "44cfe60a-9917-4967-a852-73452c90a012" (UID: "44cfe60a-9917-4967-a852-73452c90a012"). InnerVolumeSpecName "kube-api-access-nqlqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.193423 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqlqv\" (UniqueName: \"kubernetes.io/projected/44cfe60a-9917-4967-a852-73452c90a012-kube-api-access-nqlqv\") on node \"crc\" DevicePath \"\"" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.580238 4741 scope.go:117] "RemoveContainer" containerID="521442e9a2af2eeadf5adde3e17ac3c09cd84ccb90d8dc8d9e7ab16e3bd6853f" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.580483 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.624777 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Sep 29 20:28:45 crc kubenswrapper[4741]: E0929 20:28:45.625787 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44cfe60a-9917-4967-a852-73452c90a012" containerName="mariadb-client-1" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.625928 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="44cfe60a-9917-4967-a852-73452c90a012" containerName="mariadb-client-1" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.626192 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="44cfe60a-9917-4967-a852-73452c90a012" containerName="mariadb-client-1" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.627016 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.633875 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t6ql4" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.634506 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.710161 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgw8p\" (UniqueName: \"kubernetes.io/projected/a5d541d8-155f-4ed0-9136-a9f041b3db36-kube-api-access-mgw8p\") pod \"mariadb-client-4-default\" (UID: \"a5d541d8-155f-4ed0-9136-a9f041b3db36\") " pod="openstack/mariadb-client-4-default" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.812385 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgw8p\" (UniqueName: \"kubernetes.io/projected/a5d541d8-155f-4ed0-9136-a9f041b3db36-kube-api-access-mgw8p\") pod \"mariadb-client-4-default\" (UID: \"a5d541d8-155f-4ed0-9136-a9f041b3db36\") " pod="openstack/mariadb-client-4-default" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.834709 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgw8p\" (UniqueName: \"kubernetes.io/projected/a5d541d8-155f-4ed0-9136-a9f041b3db36-kube-api-access-mgw8p\") pod \"mariadb-client-4-default\" (UID: \"a5d541d8-155f-4ed0-9136-a9f041b3db36\") " pod="openstack/mariadb-client-4-default" Sep 29 20:28:45 crc kubenswrapper[4741]: I0929 20:28:45.964364 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Sep 29 20:28:46 crc kubenswrapper[4741]: I0929 20:28:46.467976 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Sep 29 20:28:46 crc kubenswrapper[4741]: W0929 20:28:46.589985 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5d541d8_155f_4ed0_9136_a9f041b3db36.slice/crio-0517bac892c0c291c481d1fbfcc9c8571b79ddeffc6345c9e2e2c8f0b21b478f WatchSource:0}: Error finding container 0517bac892c0c291c481d1fbfcc9c8571b79ddeffc6345c9e2e2c8f0b21b478f: Status 404 returned error can't find the container with id 0517bac892c0c291c481d1fbfcc9c8571b79ddeffc6345c9e2e2c8f0b21b478f Sep 29 20:28:47 crc kubenswrapper[4741]: I0929 20:28:47.097147 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44cfe60a-9917-4967-a852-73452c90a012" path="/var/lib/kubelet/pods/44cfe60a-9917-4967-a852-73452c90a012/volumes" Sep 29 20:28:47 crc kubenswrapper[4741]: I0929 20:28:47.599169 4741 generic.go:334] "Generic (PLEG): container finished" podID="a5d541d8-155f-4ed0-9136-a9f041b3db36" containerID="a34dbbf9df8af1781b500f940963272c02126174352c3773891f2327fa278680" exitCode=0 Sep 29 20:28:47 crc kubenswrapper[4741]: I0929 20:28:47.599223 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"a5d541d8-155f-4ed0-9136-a9f041b3db36","Type":"ContainerDied","Data":"a34dbbf9df8af1781b500f940963272c02126174352c3773891f2327fa278680"} Sep 29 20:28:47 crc kubenswrapper[4741]: I0929 20:28:47.599274 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"a5d541d8-155f-4ed0-9136-a9f041b3db36","Type":"ContainerStarted","Data":"0517bac892c0c291c481d1fbfcc9c8571b79ddeffc6345c9e2e2c8f0b21b478f"} Sep 29 20:28:48 crc kubenswrapper[4741]: I0929 20:28:48.979035 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Sep 29 20:28:48 crc kubenswrapper[4741]: I0929 20:28:48.993939 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_a5d541d8-155f-4ed0-9136-a9f041b3db36/mariadb-client-4-default/0.log" Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.022405 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.028331 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.059071 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgw8p\" (UniqueName: \"kubernetes.io/projected/a5d541d8-155f-4ed0-9136-a9f041b3db36-kube-api-access-mgw8p\") pod \"a5d541d8-155f-4ed0-9136-a9f041b3db36\" (UID: \"a5d541d8-155f-4ed0-9136-a9f041b3db36\") " Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.064054 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5d541d8-155f-4ed0-9136-a9f041b3db36-kube-api-access-mgw8p" (OuterVolumeSpecName: "kube-api-access-mgw8p") pod "a5d541d8-155f-4ed0-9136-a9f041b3db36" (UID: "a5d541d8-155f-4ed0-9136-a9f041b3db36"). InnerVolumeSpecName "kube-api-access-mgw8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.106641 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5d541d8-155f-4ed0-9136-a9f041b3db36" path="/var/lib/kubelet/pods/a5d541d8-155f-4ed0-9136-a9f041b3db36/volumes" Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.160761 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgw8p\" (UniqueName: \"kubernetes.io/projected/a5d541d8-155f-4ed0-9136-a9f041b3db36-kube-api-access-mgw8p\") on node \"crc\" DevicePath \"\"" Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.616776 4741 scope.go:117] "RemoveContainer" containerID="a34dbbf9df8af1781b500f940963272c02126174352c3773891f2327fa278680" Sep 29 20:28:49 crc kubenswrapper[4741]: I0929 20:28:49.617174 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.086478 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:28:53 crc kubenswrapper[4741]: E0929 20:28:53.087019 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.585658 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Sep 29 20:28:53 crc kubenswrapper[4741]: E0929 20:28:53.586153 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d541d8-155f-4ed0-9136-a9f041b3db36" containerName="mariadb-client-4-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.586183 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d541d8-155f-4ed0-9136-a9f041b3db36" containerName="mariadb-client-4-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.586483 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d541d8-155f-4ed0-9136-a9f041b3db36" containerName="mariadb-client-4-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.587334 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.589342 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t6ql4" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.592749 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.637743 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bxbs\" (UniqueName: \"kubernetes.io/projected/32b36a6a-dd2b-42ea-94df-1edb3a3061f9-kube-api-access-9bxbs\") pod \"mariadb-client-5-default\" (UID: \"32b36a6a-dd2b-42ea-94df-1edb3a3061f9\") " pod="openstack/mariadb-client-5-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.739691 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bxbs\" (UniqueName: \"kubernetes.io/projected/32b36a6a-dd2b-42ea-94df-1edb3a3061f9-kube-api-access-9bxbs\") pod \"mariadb-client-5-default\" (UID: \"32b36a6a-dd2b-42ea-94df-1edb3a3061f9\") " pod="openstack/mariadb-client-5-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.759619 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bxbs\" (UniqueName: \"kubernetes.io/projected/32b36a6a-dd2b-42ea-94df-1edb3a3061f9-kube-api-access-9bxbs\") pod \"mariadb-client-5-default\" (UID: \"32b36a6a-dd2b-42ea-94df-1edb3a3061f9\") " pod="openstack/mariadb-client-5-default" Sep 29 20:28:53 crc kubenswrapper[4741]: I0929 20:28:53.910944 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Sep 29 20:28:54 crc kubenswrapper[4741]: I0929 20:28:54.425040 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Sep 29 20:28:54 crc kubenswrapper[4741]: I0929 20:28:54.670862 4741 generic.go:334] "Generic (PLEG): container finished" podID="32b36a6a-dd2b-42ea-94df-1edb3a3061f9" containerID="ab3382abf320ff83b6d2751c76f189021d8921fca2886218c2d19a6dcb670fb3" exitCode=0 Sep 29 20:28:54 crc kubenswrapper[4741]: I0929 20:28:54.671115 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"32b36a6a-dd2b-42ea-94df-1edb3a3061f9","Type":"ContainerDied","Data":"ab3382abf320ff83b6d2751c76f189021d8921fca2886218c2d19a6dcb670fb3"} Sep 29 20:28:54 crc kubenswrapper[4741]: I0929 20:28:54.671217 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"32b36a6a-dd2b-42ea-94df-1edb3a3061f9","Type":"ContainerStarted","Data":"6177d7f8489ec39fd322b697777a2d43b38da8fc4b2a71c5ed86450530b27146"} Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.062121 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.087826 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_32b36a6a-dd2b-42ea-94df-1edb3a3061f9/mariadb-client-5-default/0.log" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.116113 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.124761 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.185555 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bxbs\" (UniqueName: \"kubernetes.io/projected/32b36a6a-dd2b-42ea-94df-1edb3a3061f9-kube-api-access-9bxbs\") pod \"32b36a6a-dd2b-42ea-94df-1edb3a3061f9\" (UID: \"32b36a6a-dd2b-42ea-94df-1edb3a3061f9\") " Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.192846 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b36a6a-dd2b-42ea-94df-1edb3a3061f9-kube-api-access-9bxbs" (OuterVolumeSpecName: "kube-api-access-9bxbs") pod "32b36a6a-dd2b-42ea-94df-1edb3a3061f9" (UID: "32b36a6a-dd2b-42ea-94df-1edb3a3061f9"). InnerVolumeSpecName "kube-api-access-9bxbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.288456 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bxbs\" (UniqueName: \"kubernetes.io/projected/32b36a6a-dd2b-42ea-94df-1edb3a3061f9-kube-api-access-9bxbs\") on node \"crc\" DevicePath \"\"" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.299733 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Sep 29 20:28:56 crc kubenswrapper[4741]: E0929 20:28:56.300137 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b36a6a-dd2b-42ea-94df-1edb3a3061f9" containerName="mariadb-client-5-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.300167 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b36a6a-dd2b-42ea-94df-1edb3a3061f9" containerName="mariadb-client-5-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.300381 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b36a6a-dd2b-42ea-94df-1edb3a3061f9" containerName="mariadb-client-5-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.301012 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.305746 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.389371 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jrml\" (UniqueName: \"kubernetes.io/projected/0a493911-74f7-44c0-9342-60b53541cdac-kube-api-access-4jrml\") pod \"mariadb-client-6-default\" (UID: \"0a493911-74f7-44c0-9342-60b53541cdac\") " pod="openstack/mariadb-client-6-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.491412 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jrml\" (UniqueName: \"kubernetes.io/projected/0a493911-74f7-44c0-9342-60b53541cdac-kube-api-access-4jrml\") pod \"mariadb-client-6-default\" (UID: \"0a493911-74f7-44c0-9342-60b53541cdac\") " pod="openstack/mariadb-client-6-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.513241 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jrml\" (UniqueName: \"kubernetes.io/projected/0a493911-74f7-44c0-9342-60b53541cdac-kube-api-access-4jrml\") pod \"mariadb-client-6-default\" (UID: \"0a493911-74f7-44c0-9342-60b53541cdac\") " pod="openstack/mariadb-client-6-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.627680 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.694239 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6177d7f8489ec39fd322b697777a2d43b38da8fc4b2a71c5ed86450530b27146" Sep 29 20:28:56 crc kubenswrapper[4741]: I0929 20:28:56.694310 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Sep 29 20:28:57 crc kubenswrapper[4741]: I0929 20:28:57.095825 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b36a6a-dd2b-42ea-94df-1edb3a3061f9" path="/var/lib/kubelet/pods/32b36a6a-dd2b-42ea-94df-1edb3a3061f9/volumes" Sep 29 20:28:57 crc kubenswrapper[4741]: I0929 20:28:57.189508 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Sep 29 20:28:57 crc kubenswrapper[4741]: I0929 20:28:57.704465 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0a493911-74f7-44c0-9342-60b53541cdac","Type":"ContainerStarted","Data":"01569dac36f48c73cc9779cc1e18fcabaa638c439972e50762b5dd3e7f17adbe"} Sep 29 20:28:58 crc kubenswrapper[4741]: I0929 20:28:58.714558 4741 generic.go:334] "Generic (PLEG): container finished" podID="0a493911-74f7-44c0-9342-60b53541cdac" containerID="9dbca03c3d04d4330f9bb1ce6a2d03ad08690e23e6f52a9afa980302611c11bc" exitCode=0 Sep 29 20:28:58 crc kubenswrapper[4741]: I0929 20:28:58.714773 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0a493911-74f7-44c0-9342-60b53541cdac","Type":"ContainerDied","Data":"9dbca03c3d04d4330f9bb1ce6a2d03ad08690e23e6f52a9afa980302611c11bc"} Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.144673 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.200489 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_0a493911-74f7-44c0-9342-60b53541cdac/mariadb-client-6-default/0.log" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.223593 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.233727 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.250523 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jrml\" (UniqueName: \"kubernetes.io/projected/0a493911-74f7-44c0-9342-60b53541cdac-kube-api-access-4jrml\") pod \"0a493911-74f7-44c0-9342-60b53541cdac\" (UID: \"0a493911-74f7-44c0-9342-60b53541cdac\") " Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.256582 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a493911-74f7-44c0-9342-60b53541cdac-kube-api-access-4jrml" (OuterVolumeSpecName: "kube-api-access-4jrml") pod "0a493911-74f7-44c0-9342-60b53541cdac" (UID: "0a493911-74f7-44c0-9342-60b53541cdac"). InnerVolumeSpecName "kube-api-access-4jrml". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.352680 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jrml\" (UniqueName: \"kubernetes.io/projected/0a493911-74f7-44c0-9342-60b53541cdac-kube-api-access-4jrml\") on node \"crc\" DevicePath \"\"" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.416046 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Sep 29 20:29:00 crc kubenswrapper[4741]: E0929 20:29:00.416752 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a493911-74f7-44c0-9342-60b53541cdac" containerName="mariadb-client-6-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.416800 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a493911-74f7-44c0-9342-60b53541cdac" containerName="mariadb-client-6-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.417226 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a493911-74f7-44c0-9342-60b53541cdac" containerName="mariadb-client-6-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.418580 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.424522 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.454902 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5jht\" (UniqueName: \"kubernetes.io/projected/c361e252-0429-4d68-ac73-0b8c9cdfec4a-kube-api-access-h5jht\") pod \"mariadb-client-7-default\" (UID: \"c361e252-0429-4d68-ac73-0b8c9cdfec4a\") " pod="openstack/mariadb-client-7-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.557070 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5jht\" (UniqueName: \"kubernetes.io/projected/c361e252-0429-4d68-ac73-0b8c9cdfec4a-kube-api-access-h5jht\") pod \"mariadb-client-7-default\" (UID: \"c361e252-0429-4d68-ac73-0b8c9cdfec4a\") " pod="openstack/mariadb-client-7-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.583642 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5jht\" (UniqueName: \"kubernetes.io/projected/c361e252-0429-4d68-ac73-0b8c9cdfec4a-kube-api-access-h5jht\") pod \"mariadb-client-7-default\" (UID: \"c361e252-0429-4d68-ac73-0b8c9cdfec4a\") " pod="openstack/mariadb-client-7-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.731559 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01569dac36f48c73cc9779cc1e18fcabaa638c439972e50762b5dd3e7f17adbe" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.731643 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Sep 29 20:29:00 crc kubenswrapper[4741]: I0929 20:29:00.747945 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Sep 29 20:29:01 crc kubenswrapper[4741]: I0929 20:29:01.040053 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Sep 29 20:29:01 crc kubenswrapper[4741]: W0929 20:29:01.043443 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc361e252_0429_4d68_ac73_0b8c9cdfec4a.slice/crio-77bc601391fc24dfcbbe0aac07e26bf2a1cb8f5e1dd816226310b738aabe3cc5 WatchSource:0}: Error finding container 77bc601391fc24dfcbbe0aac07e26bf2a1cb8f5e1dd816226310b738aabe3cc5: Status 404 returned error can't find the container with id 77bc601391fc24dfcbbe0aac07e26bf2a1cb8f5e1dd816226310b738aabe3cc5 Sep 29 20:29:01 crc kubenswrapper[4741]: I0929 20:29:01.095015 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a493911-74f7-44c0-9342-60b53541cdac" path="/var/lib/kubelet/pods/0a493911-74f7-44c0-9342-60b53541cdac/volumes" Sep 29 20:29:01 crc kubenswrapper[4741]: I0929 20:29:01.742128 4741 generic.go:334] "Generic (PLEG): container finished" podID="c361e252-0429-4d68-ac73-0b8c9cdfec4a" containerID="3a7a00fdb32edde507bc32b885dd92dec63a439b9f4991531215115bf39a7bce" exitCode=0 Sep 29 20:29:01 crc kubenswrapper[4741]: I0929 20:29:01.742165 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c361e252-0429-4d68-ac73-0b8c9cdfec4a","Type":"ContainerDied","Data":"3a7a00fdb32edde507bc32b885dd92dec63a439b9f4991531215115bf39a7bce"} Sep 29 20:29:01 crc kubenswrapper[4741]: I0929 20:29:01.742191 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c361e252-0429-4d68-ac73-0b8c9cdfec4a","Type":"ContainerStarted","Data":"77bc601391fc24dfcbbe0aac07e26bf2a1cb8f5e1dd816226310b738aabe3cc5"} Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.138077 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.156255 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_c361e252-0429-4d68-ac73-0b8c9cdfec4a/mariadb-client-7-default/0.log" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.185909 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.196585 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.206224 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5jht\" (UniqueName: \"kubernetes.io/projected/c361e252-0429-4d68-ac73-0b8c9cdfec4a-kube-api-access-h5jht\") pod \"c361e252-0429-4d68-ac73-0b8c9cdfec4a\" (UID: \"c361e252-0429-4d68-ac73-0b8c9cdfec4a\") " Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.213521 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c361e252-0429-4d68-ac73-0b8c9cdfec4a-kube-api-access-h5jht" (OuterVolumeSpecName: "kube-api-access-h5jht") pod "c361e252-0429-4d68-ac73-0b8c9cdfec4a" (UID: "c361e252-0429-4d68-ac73-0b8c9cdfec4a"). InnerVolumeSpecName "kube-api-access-h5jht". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.307887 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5jht\" (UniqueName: \"kubernetes.io/projected/c361e252-0429-4d68-ac73-0b8c9cdfec4a-kube-api-access-h5jht\") on node \"crc\" DevicePath \"\"" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.342293 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Sep 29 20:29:03 crc kubenswrapper[4741]: E0929 20:29:03.342721 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c361e252-0429-4d68-ac73-0b8c9cdfec4a" containerName="mariadb-client-7-default" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.342746 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c361e252-0429-4d68-ac73-0b8c9cdfec4a" containerName="mariadb-client-7-default" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.342994 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c361e252-0429-4d68-ac73-0b8c9cdfec4a" containerName="mariadb-client-7-default" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.343656 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.352812 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.409778 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8snrz\" (UniqueName: \"kubernetes.io/projected/f6d63e87-9e77-4b58-8e57-5e4a94c8721f-kube-api-access-8snrz\") pod \"mariadb-client-2\" (UID: \"f6d63e87-9e77-4b58-8e57-5e4a94c8721f\") " pod="openstack/mariadb-client-2" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.511289 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8snrz\" (UniqueName: \"kubernetes.io/projected/f6d63e87-9e77-4b58-8e57-5e4a94c8721f-kube-api-access-8snrz\") pod \"mariadb-client-2\" (UID: \"f6d63e87-9e77-4b58-8e57-5e4a94c8721f\") " pod="openstack/mariadb-client-2" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.531959 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8snrz\" (UniqueName: \"kubernetes.io/projected/f6d63e87-9e77-4b58-8e57-5e4a94c8721f-kube-api-access-8snrz\") pod \"mariadb-client-2\" (UID: \"f6d63e87-9e77-4b58-8e57-5e4a94c8721f\") " pod="openstack/mariadb-client-2" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.677532 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.769771 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77bc601391fc24dfcbbe0aac07e26bf2a1cb8f5e1dd816226310b738aabe3cc5" Sep 29 20:29:03 crc kubenswrapper[4741]: I0929 20:29:03.769845 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Sep 29 20:29:04 crc kubenswrapper[4741]: I0929 20:29:04.171520 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Sep 29 20:29:04 crc kubenswrapper[4741]: I0929 20:29:04.777666 4741 generic.go:334] "Generic (PLEG): container finished" podID="f6d63e87-9e77-4b58-8e57-5e4a94c8721f" containerID="12501422c754abfa4c4bc19b10a7120f658748f1a3abcbf9334cf7321e23e133" exitCode=0 Sep 29 20:29:04 crc kubenswrapper[4741]: I0929 20:29:04.777718 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"f6d63e87-9e77-4b58-8e57-5e4a94c8721f","Type":"ContainerDied","Data":"12501422c754abfa4c4bc19b10a7120f658748f1a3abcbf9334cf7321e23e133"} Sep 29 20:29:04 crc kubenswrapper[4741]: I0929 20:29:04.777756 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"f6d63e87-9e77-4b58-8e57-5e4a94c8721f","Type":"ContainerStarted","Data":"2185b9f42c5038489041e170d20800be7bbdf3b9072de71a2fbfb0223ccca367"} Sep 29 20:29:05 crc kubenswrapper[4741]: I0929 20:29:05.086280 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:29:05 crc kubenswrapper[4741]: E0929 20:29:05.086546 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:29:05 crc kubenswrapper[4741]: I0929 20:29:05.094640 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c361e252-0429-4d68-ac73-0b8c9cdfec4a" path="/var/lib/kubelet/pods/c361e252-0429-4d68-ac73-0b8c9cdfec4a/volumes" Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.171572 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.189030 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_f6d63e87-9e77-4b58-8e57-5e4a94c8721f/mariadb-client-2/0.log" Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.219509 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.228724 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.247860 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8snrz\" (UniqueName: \"kubernetes.io/projected/f6d63e87-9e77-4b58-8e57-5e4a94c8721f-kube-api-access-8snrz\") pod \"f6d63e87-9e77-4b58-8e57-5e4a94c8721f\" (UID: \"f6d63e87-9e77-4b58-8e57-5e4a94c8721f\") " Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.255127 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d63e87-9e77-4b58-8e57-5e4a94c8721f-kube-api-access-8snrz" (OuterVolumeSpecName: "kube-api-access-8snrz") pod "f6d63e87-9e77-4b58-8e57-5e4a94c8721f" (UID: "f6d63e87-9e77-4b58-8e57-5e4a94c8721f"). InnerVolumeSpecName "kube-api-access-8snrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.349672 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8snrz\" (UniqueName: \"kubernetes.io/projected/f6d63e87-9e77-4b58-8e57-5e4a94c8721f-kube-api-access-8snrz\") on node \"crc\" DevicePath \"\"" Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.793874 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2185b9f42c5038489041e170d20800be7bbdf3b9072de71a2fbfb0223ccca367" Sep 29 20:29:06 crc kubenswrapper[4741]: I0929 20:29:06.793929 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Sep 29 20:29:07 crc kubenswrapper[4741]: I0929 20:29:07.099048 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d63e87-9e77-4b58-8e57-5e4a94c8721f" path="/var/lib/kubelet/pods/f6d63e87-9e77-4b58-8e57-5e4a94c8721f/volumes" Sep 29 20:29:19 crc kubenswrapper[4741]: I0929 20:29:19.085670 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:29:19 crc kubenswrapper[4741]: E0929 20:29:19.086697 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:29:26 crc kubenswrapper[4741]: I0929 20:29:26.221059 4741 scope.go:117] "RemoveContainer" containerID="32aae837832bdd03c822cf2fb7c9f7642fcb0bd2ac8435ebb3794c0b27312e5e" Sep 29 20:29:33 crc kubenswrapper[4741]: I0929 20:29:33.086110 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:29:34 crc kubenswrapper[4741]: I0929 20:29:34.014284 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"412450edb495dc305601b3802bc1fa8b0e748429e9746636a17baa95e321745c"} Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.178240 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn"] Sep 29 20:30:00 crc kubenswrapper[4741]: E0929 20:30:00.179438 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d63e87-9e77-4b58-8e57-5e4a94c8721f" containerName="mariadb-client-2" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.179462 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d63e87-9e77-4b58-8e57-5e4a94c8721f" containerName="mariadb-client-2" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.179892 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d63e87-9e77-4b58-8e57-5e4a94c8721f" containerName="mariadb-client-2" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.181687 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.193082 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn"] Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.194608 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.194923 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.368458 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75859263-068b-4079-93f2-f4015b2c5034-secret-volume\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.368531 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75859263-068b-4079-93f2-f4015b2c5034-config-volume\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.368599 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn6cc\" (UniqueName: \"kubernetes.io/projected/75859263-068b-4079-93f2-f4015b2c5034-kube-api-access-zn6cc\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.469931 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn6cc\" (UniqueName: \"kubernetes.io/projected/75859263-068b-4079-93f2-f4015b2c5034-kube-api-access-zn6cc\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.470434 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75859263-068b-4079-93f2-f4015b2c5034-secret-volume\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.471409 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75859263-068b-4079-93f2-f4015b2c5034-config-volume\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.472200 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75859263-068b-4079-93f2-f4015b2c5034-config-volume\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.488294 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn6cc\" (UniqueName: \"kubernetes.io/projected/75859263-068b-4079-93f2-f4015b2c5034-kube-api-access-zn6cc\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.492845 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75859263-068b-4079-93f2-f4015b2c5034-secret-volume\") pod \"collect-profiles-29319630-cgswn\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.521678 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:00 crc kubenswrapper[4741]: I0929 20:30:00.948938 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn"] Sep 29 20:30:00 crc kubenswrapper[4741]: W0929 20:30:00.964693 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75859263_068b_4079_93f2_f4015b2c5034.slice/crio-d3524cf53f9994d8f21efb9c9751ee0ac0b3451713ac3bfac49978815e42d991 WatchSource:0}: Error finding container d3524cf53f9994d8f21efb9c9751ee0ac0b3451713ac3bfac49978815e42d991: Status 404 returned error can't find the container with id d3524cf53f9994d8f21efb9c9751ee0ac0b3451713ac3bfac49978815e42d991 Sep 29 20:30:01 crc kubenswrapper[4741]: I0929 20:30:01.245314 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" event={"ID":"75859263-068b-4079-93f2-f4015b2c5034","Type":"ContainerStarted","Data":"decfd7ffa52e99c7d110c5a4ad978904e36dd3d76f2e1bb01f3adac1eb8d3cd8"} Sep 29 20:30:01 crc kubenswrapper[4741]: I0929 20:30:01.245594 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" event={"ID":"75859263-068b-4079-93f2-f4015b2c5034","Type":"ContainerStarted","Data":"d3524cf53f9994d8f21efb9c9751ee0ac0b3451713ac3bfac49978815e42d991"} Sep 29 20:30:01 crc kubenswrapper[4741]: I0929 20:30:01.266024 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" podStartSLOduration=1.266000239 podStartE2EDuration="1.266000239s" podCreationTimestamp="2025-09-29 20:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:30:01.263711667 +0000 UTC m=+4842.911501009" watchObservedRunningTime="2025-09-29 20:30:01.266000239 +0000 UTC m=+4842.913789571" Sep 29 20:30:02 crc kubenswrapper[4741]: I0929 20:30:02.254000 4741 generic.go:334] "Generic (PLEG): container finished" podID="75859263-068b-4079-93f2-f4015b2c5034" containerID="decfd7ffa52e99c7d110c5a4ad978904e36dd3d76f2e1bb01f3adac1eb8d3cd8" exitCode=0 Sep 29 20:30:02 crc kubenswrapper[4741]: I0929 20:30:02.254045 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" event={"ID":"75859263-068b-4079-93f2-f4015b2c5034","Type":"ContainerDied","Data":"decfd7ffa52e99c7d110c5a4ad978904e36dd3d76f2e1bb01f3adac1eb8d3cd8"} Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.527186 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.723273 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75859263-068b-4079-93f2-f4015b2c5034-config-volume\") pod \"75859263-068b-4079-93f2-f4015b2c5034\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.723364 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75859263-068b-4079-93f2-f4015b2c5034-secret-volume\") pod \"75859263-068b-4079-93f2-f4015b2c5034\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.723440 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn6cc\" (UniqueName: \"kubernetes.io/projected/75859263-068b-4079-93f2-f4015b2c5034-kube-api-access-zn6cc\") pod \"75859263-068b-4079-93f2-f4015b2c5034\" (UID: \"75859263-068b-4079-93f2-f4015b2c5034\") " Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.724834 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75859263-068b-4079-93f2-f4015b2c5034-config-volume" (OuterVolumeSpecName: "config-volume") pod "75859263-068b-4079-93f2-f4015b2c5034" (UID: "75859263-068b-4079-93f2-f4015b2c5034"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.729996 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75859263-068b-4079-93f2-f4015b2c5034-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "75859263-068b-4079-93f2-f4015b2c5034" (UID: "75859263-068b-4079-93f2-f4015b2c5034"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.730570 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75859263-068b-4079-93f2-f4015b2c5034-kube-api-access-zn6cc" (OuterVolumeSpecName: "kube-api-access-zn6cc") pod "75859263-068b-4079-93f2-f4015b2c5034" (UID: "75859263-068b-4079-93f2-f4015b2c5034"). InnerVolumeSpecName "kube-api-access-zn6cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.825024 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75859263-068b-4079-93f2-f4015b2c5034-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.825073 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75859263-068b-4079-93f2-f4015b2c5034-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:30:03 crc kubenswrapper[4741]: I0929 20:30:03.825094 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn6cc\" (UniqueName: \"kubernetes.io/projected/75859263-068b-4079-93f2-f4015b2c5034-kube-api-access-zn6cc\") on node \"crc\" DevicePath \"\"" Sep 29 20:30:04 crc kubenswrapper[4741]: I0929 20:30:04.269951 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" event={"ID":"75859263-068b-4079-93f2-f4015b2c5034","Type":"ContainerDied","Data":"d3524cf53f9994d8f21efb9c9751ee0ac0b3451713ac3bfac49978815e42d991"} Sep 29 20:30:04 crc kubenswrapper[4741]: I0929 20:30:04.270260 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3524cf53f9994d8f21efb9c9751ee0ac0b3451713ac3bfac49978815e42d991" Sep 29 20:30:04 crc kubenswrapper[4741]: I0929 20:30:04.270053 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn" Sep 29 20:30:04 crc kubenswrapper[4741]: I0929 20:30:04.324646 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc"] Sep 29 20:30:04 crc kubenswrapper[4741]: I0929 20:30:04.334454 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319585-v42tc"] Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.101880 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe76ac4-2e3d-4366-9ce6-075ff7eb946e" path="/var/lib/kubelet/pods/dfe76ac4-2e3d-4366-9ce6-075ff7eb946e/volumes" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.622485 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5g67f"] Sep 29 20:30:05 crc kubenswrapper[4741]: E0929 20:30:05.622838 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75859263-068b-4079-93f2-f4015b2c5034" containerName="collect-profiles" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.622853 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="75859263-068b-4079-93f2-f4015b2c5034" containerName="collect-profiles" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.623092 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="75859263-068b-4079-93f2-f4015b2c5034" containerName="collect-profiles" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.624834 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.635705 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5g67f"] Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.759826 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-utilities\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.759942 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7p9l\" (UniqueName: \"kubernetes.io/projected/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-kube-api-access-q7p9l\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.760048 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-catalog-content\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.860830 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-catalog-content\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.860896 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-utilities\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.860944 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7p9l\" (UniqueName: \"kubernetes.io/projected/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-kube-api-access-q7p9l\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.861481 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-catalog-content\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.861495 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-utilities\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.878761 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7p9l\" (UniqueName: \"kubernetes.io/projected/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-kube-api-access-q7p9l\") pod \"community-operators-5g67f\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:05 crc kubenswrapper[4741]: I0929 20:30:05.971945 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:06 crc kubenswrapper[4741]: I0929 20:30:06.212005 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5g67f"] Sep 29 20:30:06 crc kubenswrapper[4741]: I0929 20:30:06.294147 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5g67f" event={"ID":"8ddfc6bd-9b72-40aa-bd4e-af540691ad69","Type":"ContainerStarted","Data":"9c2e0836c765f4411543c38c3db6cd4648d8e9d021bfaa05a4f92ed337bfdff8"} Sep 29 20:30:07 crc kubenswrapper[4741]: I0929 20:30:07.305504 4741 generic.go:334] "Generic (PLEG): container finished" podID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerID="33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313" exitCode=0 Sep 29 20:30:07 crc kubenswrapper[4741]: I0929 20:30:07.305555 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5g67f" event={"ID":"8ddfc6bd-9b72-40aa-bd4e-af540691ad69","Type":"ContainerDied","Data":"33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313"} Sep 29 20:30:07 crc kubenswrapper[4741]: I0929 20:30:07.307829 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:30:09 crc kubenswrapper[4741]: I0929 20:30:09.339712 4741 generic.go:334] "Generic (PLEG): container finished" podID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerID="d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0" exitCode=0 Sep 29 20:30:09 crc kubenswrapper[4741]: I0929 20:30:09.340687 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5g67f" event={"ID":"8ddfc6bd-9b72-40aa-bd4e-af540691ad69","Type":"ContainerDied","Data":"d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0"} Sep 29 20:30:10 crc kubenswrapper[4741]: I0929 20:30:10.353673 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5g67f" event={"ID":"8ddfc6bd-9b72-40aa-bd4e-af540691ad69","Type":"ContainerStarted","Data":"ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12"} Sep 29 20:30:10 crc kubenswrapper[4741]: I0929 20:30:10.377783 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5g67f" podStartSLOduration=2.593286514 podStartE2EDuration="5.377753702s" podCreationTimestamp="2025-09-29 20:30:05 +0000 UTC" firstStartedPulling="2025-09-29 20:30:07.307516874 +0000 UTC m=+4848.955306206" lastFinishedPulling="2025-09-29 20:30:10.091984032 +0000 UTC m=+4851.739773394" observedRunningTime="2025-09-29 20:30:10.376349738 +0000 UTC m=+4852.024139070" watchObservedRunningTime="2025-09-29 20:30:10.377753702 +0000 UTC m=+4852.025543064" Sep 29 20:30:15 crc kubenswrapper[4741]: I0929 20:30:15.973041 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:15 crc kubenswrapper[4741]: I0929 20:30:15.973797 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:16 crc kubenswrapper[4741]: I0929 20:30:16.033559 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:16 crc kubenswrapper[4741]: I0929 20:30:16.476358 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:16 crc kubenswrapper[4741]: I0929 20:30:16.547346 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5g67f"] Sep 29 20:30:18 crc kubenswrapper[4741]: I0929 20:30:18.449234 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5g67f" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="registry-server" containerID="cri-o://ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12" gracePeriod=2 Sep 29 20:30:18 crc kubenswrapper[4741]: I0929 20:30:18.932830 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.076885 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-utilities\") pod \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.076949 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-catalog-content\") pod \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.077083 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7p9l\" (UniqueName: \"kubernetes.io/projected/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-kube-api-access-q7p9l\") pod \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\" (UID: \"8ddfc6bd-9b72-40aa-bd4e-af540691ad69\") " Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.078319 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-utilities" (OuterVolumeSpecName: "utilities") pod "8ddfc6bd-9b72-40aa-bd4e-af540691ad69" (UID: "8ddfc6bd-9b72-40aa-bd4e-af540691ad69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.090525 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-kube-api-access-q7p9l" (OuterVolumeSpecName: "kube-api-access-q7p9l") pod "8ddfc6bd-9b72-40aa-bd4e-af540691ad69" (UID: "8ddfc6bd-9b72-40aa-bd4e-af540691ad69"). InnerVolumeSpecName "kube-api-access-q7p9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.180331 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7p9l\" (UniqueName: \"kubernetes.io/projected/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-kube-api-access-q7p9l\") on node \"crc\" DevicePath \"\"" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.180381 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.182124 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ddfc6bd-9b72-40aa-bd4e-af540691ad69" (UID: "8ddfc6bd-9b72-40aa-bd4e-af540691ad69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.282930 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ddfc6bd-9b72-40aa-bd4e-af540691ad69-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.462279 4741 generic.go:334] "Generic (PLEG): container finished" podID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerID="ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12" exitCode=0 Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.462372 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5g67f" event={"ID":"8ddfc6bd-9b72-40aa-bd4e-af540691ad69","Type":"ContainerDied","Data":"ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12"} Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.462470 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5g67f" event={"ID":"8ddfc6bd-9b72-40aa-bd4e-af540691ad69","Type":"ContainerDied","Data":"9c2e0836c765f4411543c38c3db6cd4648d8e9d021bfaa05a4f92ed337bfdff8"} Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.462511 4741 scope.go:117] "RemoveContainer" containerID="ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.462525 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5g67f" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.493221 4741 scope.go:117] "RemoveContainer" containerID="d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.519068 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5g67f"] Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.533044 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5g67f"] Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.537777 4741 scope.go:117] "RemoveContainer" containerID="33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.565724 4741 scope.go:117] "RemoveContainer" containerID="ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12" Sep 29 20:30:19 crc kubenswrapper[4741]: E0929 20:30:19.566627 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12\": container with ID starting with ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12 not found: ID does not exist" containerID="ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.566708 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12"} err="failed to get container status \"ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12\": rpc error: code = NotFound desc = could not find container \"ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12\": container with ID starting with ed4aab8f2a88de540f424971fa2647cc1939b314772961d2bcfeefdc23dc1b12 not found: ID does not exist" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.566760 4741 scope.go:117] "RemoveContainer" containerID="d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0" Sep 29 20:30:19 crc kubenswrapper[4741]: E0929 20:30:19.567233 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0\": container with ID starting with d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0 not found: ID does not exist" containerID="d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.567317 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0"} err="failed to get container status \"d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0\": rpc error: code = NotFound desc = could not find container \"d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0\": container with ID starting with d1499136cbd45a62686ce820bd98efd6029b3da3e295b9fb1053d6b1a70c49c0 not found: ID does not exist" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.567382 4741 scope.go:117] "RemoveContainer" containerID="33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313" Sep 29 20:30:19 crc kubenswrapper[4741]: E0929 20:30:19.568030 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313\": container with ID starting with 33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313 not found: ID does not exist" containerID="33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313" Sep 29 20:30:19 crc kubenswrapper[4741]: I0929 20:30:19.568257 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313"} err="failed to get container status \"33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313\": rpc error: code = NotFound desc = could not find container \"33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313\": container with ID starting with 33a4dbcc4a72f7379542dca7ca4a4419a5c044ecc8bd088f10ba729eb5e87313 not found: ID does not exist" Sep 29 20:30:21 crc kubenswrapper[4741]: I0929 20:30:21.105768 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" path="/var/lib/kubelet/pods/8ddfc6bd-9b72-40aa-bd4e-af540691ad69/volumes" Sep 29 20:30:26 crc kubenswrapper[4741]: I0929 20:30:26.326959 4741 scope.go:117] "RemoveContainer" containerID="fd0c67f59704a3a436a57d65109ce936b05199316e906b69d938fd74e4e423b0" Sep 29 20:32:01 crc kubenswrapper[4741]: I0929 20:32:01.739047 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:32:01 crc kubenswrapper[4741]: I0929 20:32:01.739851 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:32:31 crc kubenswrapper[4741]: I0929 20:32:31.739073 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:32:31 crc kubenswrapper[4741]: I0929 20:32:31.739624 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.739164 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.739662 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.739700 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.740292 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"412450edb495dc305601b3802bc1fa8b0e748429e9746636a17baa95e321745c"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.740338 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://412450edb495dc305601b3802bc1fa8b0e748429e9746636a17baa95e321745c" gracePeriod=600 Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.895135 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="412450edb495dc305601b3802bc1fa8b0e748429e9746636a17baa95e321745c" exitCode=0 Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.895184 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"412450edb495dc305601b3802bc1fa8b0e748429e9746636a17baa95e321745c"} Sep 29 20:33:01 crc kubenswrapper[4741]: I0929 20:33:01.896078 4741 scope.go:117] "RemoveContainer" containerID="6712e3cba644e2ede45b9aa1e1012041f8d40a5608e304034a13a0ac10bc73a2" Sep 29 20:33:02 crc kubenswrapper[4741]: I0929 20:33:02.904264 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4"} Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.348571 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Sep 29 20:33:16 crc kubenswrapper[4741]: E0929 20:33:16.349914 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="registry-server" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.349945 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="registry-server" Sep 29 20:33:16 crc kubenswrapper[4741]: E0929 20:33:16.350005 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="extract-content" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.350028 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="extract-content" Sep 29 20:33:16 crc kubenswrapper[4741]: E0929 20:33:16.350071 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="extract-utilities" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.350086 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="extract-utilities" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.350484 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ddfc6bd-9b72-40aa-bd4e-af540691ad69" containerName="registry-server" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.351536 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.355121 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t6ql4" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.361301 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.425462 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\") pod \"mariadb-copy-data\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.425591 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnjbg\" (UniqueName: \"kubernetes.io/projected/e16f6002-f781-4228-aaa2-458b21b72250-kube-api-access-lnjbg\") pod \"mariadb-copy-data\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.527462 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnjbg\" (UniqueName: \"kubernetes.io/projected/e16f6002-f781-4228-aaa2-458b21b72250-kube-api-access-lnjbg\") pod \"mariadb-copy-data\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.527583 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\") pod \"mariadb-copy-data\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.530673 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.530721 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\") pod \"mariadb-copy-data\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9b19501fcd54e886e4807471b8f4db082779a8d111a7cbbdc60e18add6365ee5/globalmount\"" pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.546835 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnjbg\" (UniqueName: \"kubernetes.io/projected/e16f6002-f781-4228-aaa2-458b21b72250-kube-api-access-lnjbg\") pod \"mariadb-copy-data\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.562335 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\") pod \"mariadb-copy-data\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " pod="openstack/mariadb-copy-data" Sep 29 20:33:16 crc kubenswrapper[4741]: I0929 20:33:16.670945 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Sep 29 20:33:17 crc kubenswrapper[4741]: I0929 20:33:17.200580 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Sep 29 20:33:18 crc kubenswrapper[4741]: I0929 20:33:18.044119 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"e16f6002-f781-4228-aaa2-458b21b72250","Type":"ContainerStarted","Data":"092fffe8c0dcbe8b2043ef01bf0ffb3e1d9899414c05f285c2da79e8254f17d9"} Sep 29 20:33:18 crc kubenswrapper[4741]: I0929 20:33:18.044474 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"e16f6002-f781-4228-aaa2-458b21b72250","Type":"ContainerStarted","Data":"af8d87447c1239e3803c6476090dabc12c4ab76b97a01288dafe6840926dffcf"} Sep 29 20:33:18 crc kubenswrapper[4741]: I0929 20:33:18.061542 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.061524249 podStartE2EDuration="3.061524249s" podCreationTimestamp="2025-09-29 20:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:33:18.058441522 +0000 UTC m=+5039.706230864" watchObservedRunningTime="2025-09-29 20:33:18.061524249 +0000 UTC m=+5039.709313601" Sep 29 20:33:20 crc kubenswrapper[4741]: I0929 20:33:20.422737 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:20 crc kubenswrapper[4741]: I0929 20:33:20.424468 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:20 crc kubenswrapper[4741]: I0929 20:33:20.438334 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:20 crc kubenswrapper[4741]: I0929 20:33:20.501533 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nwkx\" (UniqueName: \"kubernetes.io/projected/76c9c7de-22a9-4e18-8b1c-dd23aa002065-kube-api-access-7nwkx\") pod \"mariadb-client\" (UID: \"76c9c7de-22a9-4e18-8b1c-dd23aa002065\") " pod="openstack/mariadb-client" Sep 29 20:33:20 crc kubenswrapper[4741]: I0929 20:33:20.603203 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nwkx\" (UniqueName: \"kubernetes.io/projected/76c9c7de-22a9-4e18-8b1c-dd23aa002065-kube-api-access-7nwkx\") pod \"mariadb-client\" (UID: \"76c9c7de-22a9-4e18-8b1c-dd23aa002065\") " pod="openstack/mariadb-client" Sep 29 20:33:20 crc kubenswrapper[4741]: I0929 20:33:20.622658 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nwkx\" (UniqueName: \"kubernetes.io/projected/76c9c7de-22a9-4e18-8b1c-dd23aa002065-kube-api-access-7nwkx\") pod \"mariadb-client\" (UID: \"76c9c7de-22a9-4e18-8b1c-dd23aa002065\") " pod="openstack/mariadb-client" Sep 29 20:33:20 crc kubenswrapper[4741]: I0929 20:33:20.740571 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:21 crc kubenswrapper[4741]: I0929 20:33:21.190764 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:22 crc kubenswrapper[4741]: I0929 20:33:22.078093 4741 generic.go:334] "Generic (PLEG): container finished" podID="76c9c7de-22a9-4e18-8b1c-dd23aa002065" containerID="048337c606cc646b0647f93462eb6f7d95be2a851f05d554430c7c9f4c5b9051" exitCode=0 Sep 29 20:33:22 crc kubenswrapper[4741]: I0929 20:33:22.078167 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"76c9c7de-22a9-4e18-8b1c-dd23aa002065","Type":"ContainerDied","Data":"048337c606cc646b0647f93462eb6f7d95be2a851f05d554430c7c9f4c5b9051"} Sep 29 20:33:22 crc kubenswrapper[4741]: I0929 20:33:22.078245 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"76c9c7de-22a9-4e18-8b1c-dd23aa002065","Type":"ContainerStarted","Data":"b93284d11fa0e9e2f41f22617436a92ee8964a3cda4e48d88945836d7360ee7a"} Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.499149 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.526198 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_76c9c7de-22a9-4e18-8b1c-dd23aa002065/mariadb-client/0.log" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.563978 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nwkx\" (UniqueName: \"kubernetes.io/projected/76c9c7de-22a9-4e18-8b1c-dd23aa002065-kube-api-access-7nwkx\") pod \"76c9c7de-22a9-4e18-8b1c-dd23aa002065\" (UID: \"76c9c7de-22a9-4e18-8b1c-dd23aa002065\") " Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.565663 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.573235 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76c9c7de-22a9-4e18-8b1c-dd23aa002065-kube-api-access-7nwkx" (OuterVolumeSpecName: "kube-api-access-7nwkx") pod "76c9c7de-22a9-4e18-8b1c-dd23aa002065" (UID: "76c9c7de-22a9-4e18-8b1c-dd23aa002065"). InnerVolumeSpecName "kube-api-access-7nwkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.578516 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.665202 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nwkx\" (UniqueName: \"kubernetes.io/projected/76c9c7de-22a9-4e18-8b1c-dd23aa002065-kube-api-access-7nwkx\") on node \"crc\" DevicePath \"\"" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.763551 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:23 crc kubenswrapper[4741]: E0929 20:33:23.764092 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c9c7de-22a9-4e18-8b1c-dd23aa002065" containerName="mariadb-client" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.764117 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c9c7de-22a9-4e18-8b1c-dd23aa002065" containerName="mariadb-client" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.764473 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c9c7de-22a9-4e18-8b1c-dd23aa002065" containerName="mariadb-client" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.765288 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.771123 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.871592 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdp7h\" (UniqueName: \"kubernetes.io/projected/877ed78d-534a-455d-82c8-679eb2163e69-kube-api-access-qdp7h\") pod \"mariadb-client\" (UID: \"877ed78d-534a-455d-82c8-679eb2163e69\") " pod="openstack/mariadb-client" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.973640 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdp7h\" (UniqueName: \"kubernetes.io/projected/877ed78d-534a-455d-82c8-679eb2163e69-kube-api-access-qdp7h\") pod \"mariadb-client\" (UID: \"877ed78d-534a-455d-82c8-679eb2163e69\") " pod="openstack/mariadb-client" Sep 29 20:33:23 crc kubenswrapper[4741]: I0929 20:33:23.989337 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdp7h\" (UniqueName: \"kubernetes.io/projected/877ed78d-534a-455d-82c8-679eb2163e69-kube-api-access-qdp7h\") pod \"mariadb-client\" (UID: \"877ed78d-534a-455d-82c8-679eb2163e69\") " pod="openstack/mariadb-client" Sep 29 20:33:24 crc kubenswrapper[4741]: I0929 20:33:24.090821 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:24 crc kubenswrapper[4741]: I0929 20:33:24.099045 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b93284d11fa0e9e2f41f22617436a92ee8964a3cda4e48d88945836d7360ee7a" Sep 29 20:33:24 crc kubenswrapper[4741]: I0929 20:33:24.099110 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:24 crc kubenswrapper[4741]: I0929 20:33:24.143319 4741 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="76c9c7de-22a9-4e18-8b1c-dd23aa002065" podUID="877ed78d-534a-455d-82c8-679eb2163e69" Sep 29 20:33:24 crc kubenswrapper[4741]: I0929 20:33:24.347225 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:25 crc kubenswrapper[4741]: I0929 20:33:25.104362 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76c9c7de-22a9-4e18-8b1c-dd23aa002065" path="/var/lib/kubelet/pods/76c9c7de-22a9-4e18-8b1c-dd23aa002065/volumes" Sep 29 20:33:25 crc kubenswrapper[4741]: I0929 20:33:25.109378 4741 generic.go:334] "Generic (PLEG): container finished" podID="877ed78d-534a-455d-82c8-679eb2163e69" containerID="8cf268c0530a682e7e2f6d79c1037bf32c36012baa14cdd7d5bdbe7b33d45480" exitCode=0 Sep 29 20:33:25 crc kubenswrapper[4741]: I0929 20:33:25.109491 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"877ed78d-534a-455d-82c8-679eb2163e69","Type":"ContainerDied","Data":"8cf268c0530a682e7e2f6d79c1037bf32c36012baa14cdd7d5bdbe7b33d45480"} Sep 29 20:33:25 crc kubenswrapper[4741]: I0929 20:33:25.109520 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"877ed78d-534a-455d-82c8-679eb2163e69","Type":"ContainerStarted","Data":"3b8c8e7957f2495416bfffc4b50a0846b5908bfe939056128554de275fa7eb42"} Sep 29 20:33:26 crc kubenswrapper[4741]: I0929 20:33:26.410564 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:26 crc kubenswrapper[4741]: I0929 20:33:26.429264 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_877ed78d-534a-455d-82c8-679eb2163e69/mariadb-client/0.log" Sep 29 20:33:26 crc kubenswrapper[4741]: I0929 20:33:26.452525 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:26 crc kubenswrapper[4741]: I0929 20:33:26.457320 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Sep 29 20:33:26 crc kubenswrapper[4741]: I0929 20:33:26.516436 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdp7h\" (UniqueName: \"kubernetes.io/projected/877ed78d-534a-455d-82c8-679eb2163e69-kube-api-access-qdp7h\") pod \"877ed78d-534a-455d-82c8-679eb2163e69\" (UID: \"877ed78d-534a-455d-82c8-679eb2163e69\") " Sep 29 20:33:26 crc kubenswrapper[4741]: I0929 20:33:26.521878 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/877ed78d-534a-455d-82c8-679eb2163e69-kube-api-access-qdp7h" (OuterVolumeSpecName: "kube-api-access-qdp7h") pod "877ed78d-534a-455d-82c8-679eb2163e69" (UID: "877ed78d-534a-455d-82c8-679eb2163e69"). InnerVolumeSpecName "kube-api-access-qdp7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:33:26 crc kubenswrapper[4741]: I0929 20:33:26.619005 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdp7h\" (UniqueName: \"kubernetes.io/projected/877ed78d-534a-455d-82c8-679eb2163e69-kube-api-access-qdp7h\") on node \"crc\" DevicePath \"\"" Sep 29 20:33:27 crc kubenswrapper[4741]: I0929 20:33:27.101692 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="877ed78d-534a-455d-82c8-679eb2163e69" path="/var/lib/kubelet/pods/877ed78d-534a-455d-82c8-679eb2163e69/volumes" Sep 29 20:33:27 crc kubenswrapper[4741]: I0929 20:33:27.122905 4741 scope.go:117] "RemoveContainer" containerID="8cf268c0530a682e7e2f6d79c1037bf32c36012baa14cdd7d5bdbe7b33d45480" Sep 29 20:33:27 crc kubenswrapper[4741]: I0929 20:33:27.122943 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.433847 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 20:33:58 crc kubenswrapper[4741]: E0929 20:33:58.435001 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877ed78d-534a-455d-82c8-679eb2163e69" containerName="mariadb-client" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.435029 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="877ed78d-534a-455d-82c8-679eb2163e69" containerName="mariadb-client" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.435328 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="877ed78d-534a-455d-82c8-679eb2163e69" containerName="mariadb-client" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.437022 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.439777 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.440138 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.440332 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-nf748" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.441651 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.452286 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.456731 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.460339 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.463998 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.468482 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.475086 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514110 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddc92-7548-4ed5-9aa1-703d220dee0f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514156 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnvdf\" (UniqueName: \"kubernetes.io/projected/a7bddc92-7548-4ed5-9aa1-703d220dee0f-kube-api-access-bnvdf\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514186 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7bddc92-7548-4ed5-9aa1-703d220dee0f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514211 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7bddc92-7548-4ed5-9aa1-703d220dee0f-config\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514239 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514269 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e48c5ed-877d-43d5-8a15-f47472801089-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514289 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk7fs\" (UniqueName: \"kubernetes.io/projected/2e48c5ed-877d-43d5-8a15-f47472801089-kube-api-access-pk7fs\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514313 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e48c5ed-877d-43d5-8a15-f47472801089-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514328 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e48c5ed-877d-43d5-8a15-f47472801089-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514347 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7bddc92-7548-4ed5-9aa1-703d220dee0f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514366 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e48c5ed-877d-43d5-8a15-f47472801089-config\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.514407 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616561 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/77eb4b78-487a-4078-91b3-98172b274a9f-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616645 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddc92-7548-4ed5-9aa1-703d220dee0f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616671 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnvdf\" (UniqueName: \"kubernetes.io/projected/a7bddc92-7548-4ed5-9aa1-703d220dee0f-kube-api-access-bnvdf\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616700 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7bddc92-7548-4ed5-9aa1-703d220dee0f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616732 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7bddc92-7548-4ed5-9aa1-703d220dee0f-config\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616782 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-380e33f8-a4b7-481f-9701-862624a4bad2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-380e33f8-a4b7-481f-9701-862624a4bad2\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616825 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616874 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e48c5ed-877d-43d5-8a15-f47472801089-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616909 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk7fs\" (UniqueName: \"kubernetes.io/projected/2e48c5ed-877d-43d5-8a15-f47472801089-kube-api-access-pk7fs\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616952 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e48c5ed-877d-43d5-8a15-f47472801089-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.616983 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e48c5ed-877d-43d5-8a15-f47472801089-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.617017 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdvw\" (UniqueName: \"kubernetes.io/projected/77eb4b78-487a-4078-91b3-98172b274a9f-kube-api-access-5kdvw\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.617051 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7bddc92-7548-4ed5-9aa1-703d220dee0f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.617089 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e48c5ed-877d-43d5-8a15-f47472801089-config\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.617129 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77eb4b78-487a-4078-91b3-98172b274a9f-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.617163 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77eb4b78-487a-4078-91b3-98172b274a9f-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.617201 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.617242 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77eb4b78-487a-4078-91b3-98172b274a9f-config\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.618010 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7bddc92-7548-4ed5-9aa1-703d220dee0f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.618091 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7bddc92-7548-4ed5-9aa1-703d220dee0f-config\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.618140 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e48c5ed-877d-43d5-8a15-f47472801089-config\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.618448 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e48c5ed-877d-43d5-8a15-f47472801089-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.618552 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e48c5ed-877d-43d5-8a15-f47472801089-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.619837 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7bddc92-7548-4ed5-9aa1-703d220dee0f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.619920 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.619942 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3aa706deff873eab7a949466d6d22d6cbfaf3d02f67ea3d891b8388e07c04a49/globalmount\"" pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.620670 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.620705 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5e8ba31c417f7e34a0cb5479570007d948ac073abaff0c347e8fa1f55f4610cf/globalmount\"" pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.633312 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.634758 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e48c5ed-877d-43d5-8a15-f47472801089-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.635044 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.636894 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-x6zn6" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.637446 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddc92-7548-4ed5-9aa1-703d220dee0f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.637788 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.638721 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.644979 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnvdf\" (UniqueName: \"kubernetes.io/projected/a7bddc92-7548-4ed5-9aa1-703d220dee0f-kube-api-access-bnvdf\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.654921 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.656379 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.658882 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk7fs\" (UniqueName: \"kubernetes.io/projected/2e48c5ed-877d-43d5-8a15-f47472801089-kube-api-access-pk7fs\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.661385 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.663890 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.670296 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.675417 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.689883 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9e9827d-d361-41f2-bc57-b645ebd1e738\") pod \"ovsdbserver-nb-2\" (UID: \"a7bddc92-7548-4ed5-9aa1-703d220dee0f\") " pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.708339 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5aa2ef46-8808-419d-bbb6-cc3f9640f6ec\") pod \"ovsdbserver-nb-0\" (UID: \"2e48c5ed-877d-43d5-8a15-f47472801089\") " pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720518 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3eed90c2-9391-40d3-819e-59d316fb7890\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3eed90c2-9391-40d3-819e-59d316fb7890\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720591 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28jbm\" (UniqueName: \"kubernetes.io/projected/ffa3c586-6f20-41bd-a638-69c1511d7652-kube-api-access-28jbm\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720629 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720708 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/77eb4b78-487a-4078-91b3-98172b274a9f-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720776 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00b2f37-4126-49e1-99e6-4cf54d225de1-config\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720814 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48bgb\" (UniqueName: \"kubernetes.io/projected/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-kube-api-access-48bgb\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720868 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-380e33f8-a4b7-481f-9701-862624a4bad2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-380e33f8-a4b7-481f-9701-862624a4bad2\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720901 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720938 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ffa3c586-6f20-41bd-a638-69c1511d7652-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.720976 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgh6g\" (UniqueName: \"kubernetes.io/projected/c00b2f37-4126-49e1-99e6-4cf54d225de1-kube-api-access-sgh6g\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721029 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721066 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ffa3c586-6f20-41bd-a638-69c1511d7652-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721116 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c00b2f37-4126-49e1-99e6-4cf54d225de1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721152 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdvw\" (UniqueName: \"kubernetes.io/projected/77eb4b78-487a-4078-91b3-98172b274a9f-kube-api-access-5kdvw\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721192 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c00b2f37-4126-49e1-99e6-4cf54d225de1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721229 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c00b2f37-4126-49e1-99e6-4cf54d225de1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721270 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77eb4b78-487a-4078-91b3-98172b274a9f-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721302 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77eb4b78-487a-4078-91b3-98172b274a9f-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721333 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa3c586-6f20-41bd-a638-69c1511d7652-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721415 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77eb4b78-487a-4078-91b3-98172b274a9f-config\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721462 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721537 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721586 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.721619 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffa3c586-6f20-41bd-a638-69c1511d7652-config\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.722669 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/77eb4b78-487a-4078-91b3-98172b274a9f-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.723341 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.724444 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77eb4b78-487a-4078-91b3-98172b274a9f-config\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.727660 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77eb4b78-487a-4078-91b3-98172b274a9f-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.735961 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.736009 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-380e33f8-a4b7-481f-9701-862624a4bad2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-380e33f8-a4b7-481f-9701-862624a4bad2\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f117576cc40c8847f4ba3d4dc6fc4aad125bf7bd45a810088d0100705e1c5be6/globalmount\"" pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.739778 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77eb4b78-487a-4078-91b3-98172b274a9f-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.744570 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdvw\" (UniqueName: \"kubernetes.io/projected/77eb4b78-487a-4078-91b3-98172b274a9f-kube-api-access-5kdvw\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.760321 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-380e33f8-a4b7-481f-9701-862624a4bad2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-380e33f8-a4b7-481f-9701-862624a4bad2\") pod \"ovsdbserver-nb-1\" (UID: \"77eb4b78-487a-4078-91b3-98172b274a9f\") " pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.776141 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.797714 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.807473 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.823781 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa3c586-6f20-41bd-a638-69c1511d7652-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.823837 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.823879 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.823904 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.823927 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffa3c586-6f20-41bd-a638-69c1511d7652-config\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.823958 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3eed90c2-9391-40d3-819e-59d316fb7890\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3eed90c2-9391-40d3-819e-59d316fb7890\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.823999 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28jbm\" (UniqueName: \"kubernetes.io/projected/ffa3c586-6f20-41bd-a638-69c1511d7652-kube-api-access-28jbm\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824026 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824069 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00b2f37-4126-49e1-99e6-4cf54d225de1-config\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824090 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48bgb\" (UniqueName: \"kubernetes.io/projected/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-kube-api-access-48bgb\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824122 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824145 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ffa3c586-6f20-41bd-a638-69c1511d7652-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824174 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgh6g\" (UniqueName: \"kubernetes.io/projected/c00b2f37-4126-49e1-99e6-4cf54d225de1-kube-api-access-sgh6g\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824208 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824234 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ffa3c586-6f20-41bd-a638-69c1511d7652-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824280 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c00b2f37-4126-49e1-99e6-4cf54d225de1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824310 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c00b2f37-4126-49e1-99e6-4cf54d225de1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.824333 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c00b2f37-4126-49e1-99e6-4cf54d225de1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.825544 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.825575 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.825853 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.826383 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c00b2f37-4126-49e1-99e6-4cf54d225de1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.826511 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffa3c586-6f20-41bd-a638-69c1511d7652-config\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.826710 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ffa3c586-6f20-41bd-a638-69c1511d7652-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.827117 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c00b2f37-4126-49e1-99e6-4cf54d225de1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.827182 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00b2f37-4126-49e1-99e6-4cf54d225de1-config\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.827728 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.828492 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ffa3c586-6f20-41bd-a638-69c1511d7652-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.828564 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.828654 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3eed90c2-9391-40d3-819e-59d316fb7890\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3eed90c2-9391-40d3-819e-59d316fb7890\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8bb0c9d5c22e12e2377bb460ba3c0e07094bd6bf7fff3a2935d30990033802cc/globalmount\"" pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.838266 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.838989 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8283aa2cec54ef81d007e31feef70cd0620262af08c76f9ce9394ed26e4ac0ff/globalmount\"" pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.838893 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c00b2f37-4126-49e1-99e6-4cf54d225de1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.839571 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.839695 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2f26d92ab2779f84597576ecd725efecf00bf8ae2c7f656d897cd71efbc80256/globalmount\"" pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.840477 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48bgb\" (UniqueName: \"kubernetes.io/projected/e5583f47-e04c-4bc3-8614-4c2d67f61a8b-kube-api-access-48bgb\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.844369 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa3c586-6f20-41bd-a638-69c1511d7652-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.848581 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28jbm\" (UniqueName: \"kubernetes.io/projected/ffa3c586-6f20-41bd-a638-69c1511d7652-kube-api-access-28jbm\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.866255 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3eed90c2-9391-40d3-819e-59d316fb7890\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3eed90c2-9391-40d3-819e-59d316fb7890\") pod \"ovsdbserver-sb-1\" (UID: \"e5583f47-e04c-4bc3-8614-4c2d67f61a8b\") " pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.871589 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgh6g\" (UniqueName: \"kubernetes.io/projected/c00b2f37-4126-49e1-99e6-4cf54d225de1-kube-api-access-sgh6g\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.877665 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1112b362-e4be-42f7-b61c-b97c5fbc12d2\") pod \"ovsdbserver-sb-0\" (UID: \"c00b2f37-4126-49e1-99e6-4cf54d225de1\") " pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:58 crc kubenswrapper[4741]: I0929 20:33:58.887478 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbcddef2-8f8d-427a-a050-bb5b2a6e1c53\") pod \"ovsdbserver-sb-2\" (UID: \"ffa3c586-6f20-41bd-a638-69c1511d7652\") " pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.124120 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.156415 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.172006 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.377273 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.402166 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2e48c5ed-877d-43d5-8a15-f47472801089","Type":"ContainerStarted","Data":"7a5f65953d501e004de6256d2e20bf3c5c9331beaa205c455d67758f1b38e8ee"} Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.467470 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.702078 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Sep 29 20:33:59 crc kubenswrapper[4741]: I0929 20:33:59.802338 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 20:33:59 crc kubenswrapper[4741]: W0929 20:33:59.811343 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc00b2f37_4126_49e1_99e6_4cf54d225de1.slice/crio-f63082cc7180e2179d5ae763a4df2f109e6289e2e3bba91142e97d2b632301ff WatchSource:0}: Error finding container f63082cc7180e2179d5ae763a4df2f109e6289e2e3bba91142e97d2b632301ff: Status 404 returned error can't find the container with id f63082cc7180e2179d5ae763a4df2f109e6289e2e3bba91142e97d2b632301ff Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.083468 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Sep 29 20:34:00 crc kubenswrapper[4741]: W0929 20:34:00.089971 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7bddc92_7548_4ed5_9aa1_703d220dee0f.slice/crio-81b7c403127d73614b2a9ca916f2faa1fd1b7bc79ec46322ab3117b906d8102e WatchSource:0}: Error finding container 81b7c403127d73614b2a9ca916f2faa1fd1b7bc79ec46322ab3117b906d8102e: Status 404 returned error can't find the container with id 81b7c403127d73614b2a9ca916f2faa1fd1b7bc79ec46322ab3117b906d8102e Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.412981 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"77eb4b78-487a-4078-91b3-98172b274a9f","Type":"ContainerStarted","Data":"90fa08dcd24ab9a00863a4840b0e851f76e1f20fe7d80332490117f06a86bdbe"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.413024 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"77eb4b78-487a-4078-91b3-98172b274a9f","Type":"ContainerStarted","Data":"a88e1a38662e0430ca59d2a2ce43989b72470e1fa25709553e30e59628b86247"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.413036 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"77eb4b78-487a-4078-91b3-98172b274a9f","Type":"ContainerStarted","Data":"4226178808d8329387c39b351f8f9ccb28a484dd5ec7e2008ac02a506348321b"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.417226 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2e48c5ed-877d-43d5-8a15-f47472801089","Type":"ContainerStarted","Data":"e1c8a41fdfecd2362d64403c3f2e7a1c8750ad60b1aa8707c77670372d2456a3"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.417257 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2e48c5ed-877d-43d5-8a15-f47472801089","Type":"ContainerStarted","Data":"b462e5aa894c0ba87472201bef83957b5917a23d806fa480b9832a19c5e894e0"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.419916 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a7bddc92-7548-4ed5-9aa1-703d220dee0f","Type":"ContainerStarted","Data":"19e8817af68b4cd3ff1ed7c7bc212108bbdcbaf77bef648b89b9a5a768d12447"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.419961 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a7bddc92-7548-4ed5-9aa1-703d220dee0f","Type":"ContainerStarted","Data":"3195e96c0366457b6d4eeb78d6531c5c57ea16dbee6a8ebcc1a39b2fe9ede68d"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.419972 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a7bddc92-7548-4ed5-9aa1-703d220dee0f","Type":"ContainerStarted","Data":"81b7c403127d73614b2a9ca916f2faa1fd1b7bc79ec46322ab3117b906d8102e"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.422220 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c00b2f37-4126-49e1-99e6-4cf54d225de1","Type":"ContainerStarted","Data":"a0e8c259c428e8e28f0e5cd0f7523daaa398d80dc917fae640036b4a1acc1aed"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.422257 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c00b2f37-4126-49e1-99e6-4cf54d225de1","Type":"ContainerStarted","Data":"8dce217095343cad71ef2992b372d965121c510c50a7baef4e5f738cbe6347e7"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.422267 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c00b2f37-4126-49e1-99e6-4cf54d225de1","Type":"ContainerStarted","Data":"f63082cc7180e2179d5ae763a4df2f109e6289e2e3bba91142e97d2b632301ff"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.424041 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"ffa3c586-6f20-41bd-a638-69c1511d7652","Type":"ContainerStarted","Data":"af2c14a6dc77521351446569f5553b99b474216035ab5b87230703999135538f"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.424066 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"ffa3c586-6f20-41bd-a638-69c1511d7652","Type":"ContainerStarted","Data":"8d6c71e86b237d0b9f89d8cd5b4ddd5ecad9e523edc16b96d1f3e2ed32ab99da"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.424075 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"ffa3c586-6f20-41bd-a638-69c1511d7652","Type":"ContainerStarted","Data":"fa0f4512c11644500595aa2db0679f0318b100d85c731b9fd415a8a6e83f7fbb"} Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.433244 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.433229711 podStartE2EDuration="3.433229711s" podCreationTimestamp="2025-09-29 20:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:00.427352896 +0000 UTC m=+5082.075142238" watchObservedRunningTime="2025-09-29 20:34:00.433229711 +0000 UTC m=+5082.081019043" Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.451871 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.451856748 podStartE2EDuration="3.451856748s" podCreationTimestamp="2025-09-29 20:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:00.450776784 +0000 UTC m=+5082.098566116" watchObservedRunningTime="2025-09-29 20:34:00.451856748 +0000 UTC m=+5082.099646080" Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.472982 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.472964173 podStartE2EDuration="3.472964173s" podCreationTimestamp="2025-09-29 20:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:00.465893541 +0000 UTC m=+5082.113682873" watchObservedRunningTime="2025-09-29 20:34:00.472964173 +0000 UTC m=+5082.120753505" Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.492710 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.492695375 podStartE2EDuration="3.492695375s" podCreationTimestamp="2025-09-29 20:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:00.489145304 +0000 UTC m=+5082.136934636" watchObservedRunningTime="2025-09-29 20:34:00.492695375 +0000 UTC m=+5082.140484707" Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.506218 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.506201261 podStartE2EDuration="3.506201261s" podCreationTimestamp="2025-09-29 20:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:00.505531871 +0000 UTC m=+5082.153321223" watchObservedRunningTime="2025-09-29 20:34:00.506201261 +0000 UTC m=+5082.153990593" Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.617893 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.997808 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dbhjw"] Sep 29 20:34:00 crc kubenswrapper[4741]: I0929 20:34:00.999474 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.014914 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbhjw"] Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.169719 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnzcr\" (UniqueName: \"kubernetes.io/projected/eef7aa30-420d-4e5f-b24b-45401527b15e-kube-api-access-pnzcr\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.170121 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-utilities\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.170153 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-catalog-content\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.271728 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnzcr\" (UniqueName: \"kubernetes.io/projected/eef7aa30-420d-4e5f-b24b-45401527b15e-kube-api-access-pnzcr\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.271865 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-utilities\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.271906 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-catalog-content\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.272431 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-catalog-content\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.273775 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-utilities\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.291535 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnzcr\" (UniqueName: \"kubernetes.io/projected/eef7aa30-420d-4e5f-b24b-45401527b15e-kube-api-access-pnzcr\") pod \"certified-operators-dbhjw\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.317329 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.454490 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5583f47-e04c-4bc3-8614-4c2d67f61a8b","Type":"ContainerStarted","Data":"bc5c6c2717a551e6b081ab2e71538c46ab91b12e9f0568c5ed4f9e59e95c8b74"} Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.454560 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5583f47-e04c-4bc3-8614-4c2d67f61a8b","Type":"ContainerStarted","Data":"2f17642b6c55689ad9185d150e82061bc089579018df728caa7811ace29a60c0"} Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.454571 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5583f47-e04c-4bc3-8614-4c2d67f61a8b","Type":"ContainerStarted","Data":"608ccfda7d122b48e703fe18b3a329adfecde7415449fbc7d4e6a07ec872d305"} Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.478620 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.478599702 podStartE2EDuration="4.478599702s" podCreationTimestamp="2025-09-29 20:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:01.476091343 +0000 UTC m=+5083.123880675" watchObservedRunningTime="2025-09-29 20:34:01.478599702 +0000 UTC m=+5083.126389034" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.776662 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.798956 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.808155 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Sep 29 20:34:01 crc kubenswrapper[4741]: W0929 20:34:01.820946 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeef7aa30_420d_4e5f_b24b_45401527b15e.slice/crio-8db659a449519b541f190311bf86475754dd27cb69d52f8cdd1f099ec7a2b43f WatchSource:0}: Error finding container 8db659a449519b541f190311bf86475754dd27cb69d52f8cdd1f099ec7a2b43f: Status 404 returned error can't find the container with id 8db659a449519b541f190311bf86475754dd27cb69d52f8cdd1f099ec7a2b43f Sep 29 20:34:01 crc kubenswrapper[4741]: I0929 20:34:01.822072 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbhjw"] Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.124352 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.157322 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.172715 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.201203 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.216310 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.463654 4741 generic.go:334] "Generic (PLEG): container finished" podID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerID="f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5" exitCode=0 Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.463705 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbhjw" event={"ID":"eef7aa30-420d-4e5f-b24b-45401527b15e","Type":"ContainerDied","Data":"f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5"} Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.463744 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbhjw" event={"ID":"eef7aa30-420d-4e5f-b24b-45401527b15e","Type":"ContainerStarted","Data":"8db659a449519b541f190311bf86475754dd27cb69d52f8cdd1f099ec7a2b43f"} Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.465331 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Sep 29 20:34:02 crc kubenswrapper[4741]: I0929 20:34:02.465360 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Sep 29 20:34:03 crc kubenswrapper[4741]: I0929 20:34:03.474463 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbhjw" event={"ID":"eef7aa30-420d-4e5f-b24b-45401527b15e","Type":"ContainerStarted","Data":"d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad"} Sep 29 20:34:03 crc kubenswrapper[4741]: I0929 20:34:03.776433 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Sep 29 20:34:03 crc kubenswrapper[4741]: I0929 20:34:03.798258 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Sep 29 20:34:03 crc kubenswrapper[4741]: I0929 20:34:03.808302 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.125713 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.216838 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.218026 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.451213 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq"] Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.454680 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.459961 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.476591 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq"] Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.483523 4741 generic.go:334] "Generic (PLEG): container finished" podID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerID="d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad" exitCode=0 Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.483674 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbhjw" event={"ID":"eef7aa30-420d-4e5f-b24b-45401527b15e","Type":"ContainerDied","Data":"d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad"} Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.629315 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.629702 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-config\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.629732 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-dns-svc\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.630682 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfhk6\" (UniqueName: \"kubernetes.io/projected/68a460d8-c74a-4c7f-a4d8-acdf5d927324-kube-api-access-tfhk6\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.731952 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfhk6\" (UniqueName: \"kubernetes.io/projected/68a460d8-c74a-4c7f-a4d8-acdf5d927324-kube-api-access-tfhk6\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.732047 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.732089 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-config\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.732115 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-dns-svc\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.733024 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-dns-svc\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.733098 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.733111 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-config\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.748927 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfhk6\" (UniqueName: \"kubernetes.io/projected/68a460d8-c74a-4c7f-a4d8-acdf5d927324-kube-api-access-tfhk6\") pod \"dnsmasq-dns-6c9c4b4dfc-bgpdq\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.770757 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.822481 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.838883 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.888427 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.895925 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Sep 29 20:34:04 crc kubenswrapper[4741]: I0929 20:34:04.954172 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.182984 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq"] Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.192996 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.220892 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c6dcccb85-lhhw9"] Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.222595 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.226891 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.233510 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c6dcccb85-lhhw9"] Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.276441 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq"] Sep 29 20:34:05 crc kubenswrapper[4741]: W0929 20:34:05.280265 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68a460d8_c74a_4c7f_a4d8_acdf5d927324.slice/crio-10410f354fecd049cba19932d6978387ede07ee054c43962e29354d2fe2d67b3 WatchSource:0}: Error finding container 10410f354fecd049cba19932d6978387ede07ee054c43962e29354d2fe2d67b3: Status 404 returned error can't find the container with id 10410f354fecd049cba19932d6978387ede07ee054c43962e29354d2fe2d67b3 Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.343319 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.343360 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.343423 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgzkc\" (UniqueName: \"kubernetes.io/projected/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-kube-api-access-lgzkc\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.343499 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-config\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.343635 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.445634 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.445676 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.445723 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgzkc\" (UniqueName: \"kubernetes.io/projected/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-kube-api-access-lgzkc\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.445754 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-config\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.445776 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.446549 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.446616 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-config\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.446654 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.447239 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.464779 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgzkc\" (UniqueName: \"kubernetes.io/projected/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-kube-api-access-lgzkc\") pod \"dnsmasq-dns-5c6dcccb85-lhhw9\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.494513 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbhjw" event={"ID":"eef7aa30-420d-4e5f-b24b-45401527b15e","Type":"ContainerStarted","Data":"af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f"} Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.496354 4741 generic.go:334] "Generic (PLEG): container finished" podID="68a460d8-c74a-4c7f-a4d8-acdf5d927324" containerID="47e4c7c1cf3c765e32c37fd1608c6b1b3a1298bf93d3fa3772b366a84c7602af" exitCode=0 Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.496505 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" event={"ID":"68a460d8-c74a-4c7f-a4d8-acdf5d927324","Type":"ContainerDied","Data":"47e4c7c1cf3c765e32c37fd1608c6b1b3a1298bf93d3fa3772b366a84c7602af"} Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.496575 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" event={"ID":"68a460d8-c74a-4c7f-a4d8-acdf5d927324","Type":"ContainerStarted","Data":"10410f354fecd049cba19932d6978387ede07ee054c43962e29354d2fe2d67b3"} Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.511890 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dbhjw" podStartSLOduration=2.981794398 podStartE2EDuration="5.511870165s" podCreationTimestamp="2025-09-29 20:34:00 +0000 UTC" firstStartedPulling="2025-09-29 20:34:02.466052347 +0000 UTC m=+5084.113841679" lastFinishedPulling="2025-09-29 20:34:04.996128114 +0000 UTC m=+5086.643917446" observedRunningTime="2025-09-29 20:34:05.509451979 +0000 UTC m=+5087.157241311" watchObservedRunningTime="2025-09-29 20:34:05.511870165 +0000 UTC m=+5087.159659497" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.542816 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.553244 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.849792 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.953737 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-config\") pod \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.954072 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-ovsdbserver-sb\") pod \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.954122 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfhk6\" (UniqueName: \"kubernetes.io/projected/68a460d8-c74a-4c7f-a4d8-acdf5d927324-kube-api-access-tfhk6\") pod \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.954170 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-dns-svc\") pod \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\" (UID: \"68a460d8-c74a-4c7f-a4d8-acdf5d927324\") " Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.962257 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a460d8-c74a-4c7f-a4d8-acdf5d927324-kube-api-access-tfhk6" (OuterVolumeSpecName: "kube-api-access-tfhk6") pod "68a460d8-c74a-4c7f-a4d8-acdf5d927324" (UID: "68a460d8-c74a-4c7f-a4d8-acdf5d927324"). InnerVolumeSpecName "kube-api-access-tfhk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.973011 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68a460d8-c74a-4c7f-a4d8-acdf5d927324" (UID: "68a460d8-c74a-4c7f-a4d8-acdf5d927324"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.977768 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-config" (OuterVolumeSpecName: "config") pod "68a460d8-c74a-4c7f-a4d8-acdf5d927324" (UID: "68a460d8-c74a-4c7f-a4d8-acdf5d927324"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:05 crc kubenswrapper[4741]: I0929 20:34:05.981639 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68a460d8-c74a-4c7f-a4d8-acdf5d927324" (UID: "68a460d8-c74a-4c7f-a4d8-acdf5d927324"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.038947 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c6dcccb85-lhhw9"] Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.055656 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.055680 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.055691 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfhk6\" (UniqueName: \"kubernetes.io/projected/68a460d8-c74a-4c7f-a4d8-acdf5d927324-kube-api-access-tfhk6\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.055699 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a460d8-c74a-4c7f-a4d8-acdf5d927324-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.505509 4741 generic.go:334] "Generic (PLEG): container finished" podID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerID="88962a05cee4ae28aa06f8f04b337dff9b4d1501c92178eb5ef81580e028b57a" exitCode=0 Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.505592 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" event={"ID":"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef","Type":"ContainerDied","Data":"88962a05cee4ae28aa06f8f04b337dff9b4d1501c92178eb5ef81580e028b57a"} Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.505619 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" event={"ID":"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef","Type":"ContainerStarted","Data":"84551570b7eeed6eb3c7e52a180c4a846f4d66e070f951686c4454212d462151"} Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.508634 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" event={"ID":"68a460d8-c74a-4c7f-a4d8-acdf5d927324","Type":"ContainerDied","Data":"10410f354fecd049cba19932d6978387ede07ee054c43962e29354d2fe2d67b3"} Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.508697 4741 scope.go:117] "RemoveContainer" containerID="47e4c7c1cf3c765e32c37fd1608c6b1b3a1298bf93d3fa3772b366a84c7602af" Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.508701 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq" Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.583435 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq"] Sep 29 20:34:06 crc kubenswrapper[4741]: I0929 20:34:06.592375 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-bgpdq"] Sep 29 20:34:07 crc kubenswrapper[4741]: I0929 20:34:07.094658 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a460d8-c74a-4c7f-a4d8-acdf5d927324" path="/var/lib/kubelet/pods/68a460d8-c74a-4c7f-a4d8-acdf5d927324/volumes" Sep 29 20:34:07 crc kubenswrapper[4741]: I0929 20:34:07.519303 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" event={"ID":"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef","Type":"ContainerStarted","Data":"06a4eefef3fe2023e87f9c4c2cf1a81bf4cfc05e5ea0cae8f9cec2a5659d5bd5"} Sep 29 20:34:07 crc kubenswrapper[4741]: I0929 20:34:07.519461 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:07 crc kubenswrapper[4741]: I0929 20:34:07.537669 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" podStartSLOduration=2.53764657 podStartE2EDuration="2.53764657s" podCreationTimestamp="2025-09-29 20:34:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:07.534679577 +0000 UTC m=+5089.182468979" watchObservedRunningTime="2025-09-29 20:34:07.53764657 +0000 UTC m=+5089.185435902" Sep 29 20:34:08 crc kubenswrapper[4741]: I0929 20:34:08.833535 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Sep 29 20:34:11 crc kubenswrapper[4741]: I0929 20:34:11.318426 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:11 crc kubenswrapper[4741]: I0929 20:34:11.318750 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:11 crc kubenswrapper[4741]: I0929 20:34:11.363184 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:11 crc kubenswrapper[4741]: I0929 20:34:11.597849 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:11 crc kubenswrapper[4741]: I0929 20:34:11.640194 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbhjw"] Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.078473 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Sep 29 20:34:12 crc kubenswrapper[4741]: E0929 20:34:12.078946 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a460d8-c74a-4c7f-a4d8-acdf5d927324" containerName="init" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.078971 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a460d8-c74a-4c7f-a4d8-acdf5d927324" containerName="init" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.079216 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a460d8-c74a-4c7f-a4d8-acdf5d927324" containerName="init" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.080057 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.085634 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.086953 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.257601 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/39cc47ab-05f6-4582-aabc-71217551efc7-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.257719 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-41089285-8dcf-44b4-8a50-2286ad169286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.257752 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n2bg\" (UniqueName: \"kubernetes.io/projected/39cc47ab-05f6-4582-aabc-71217551efc7-kube-api-access-2n2bg\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.359528 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/39cc47ab-05f6-4582-aabc-71217551efc7-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.359611 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-41089285-8dcf-44b4-8a50-2286ad169286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.359636 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n2bg\" (UniqueName: \"kubernetes.io/projected/39cc47ab-05f6-4582-aabc-71217551efc7-kube-api-access-2n2bg\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.363159 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.363186 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-41089285-8dcf-44b4-8a50-2286ad169286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d22405de26be4d8980e52590771e41cf3b39a55165ab861fabec2f77685d0943/globalmount\"" pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.364895 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/39cc47ab-05f6-4582-aabc-71217551efc7-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.379126 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n2bg\" (UniqueName: \"kubernetes.io/projected/39cc47ab-05f6-4582-aabc-71217551efc7-kube-api-access-2n2bg\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.388653 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-41089285-8dcf-44b4-8a50-2286ad169286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286\") pod \"ovn-copy-data\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.462057 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Sep 29 20:34:12 crc kubenswrapper[4741]: I0929 20:34:12.973382 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Sep 29 20:34:13 crc kubenswrapper[4741]: I0929 20:34:13.570400 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"39cc47ab-05f6-4582-aabc-71217551efc7","Type":"ContainerStarted","Data":"c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4"} Sep 29 20:34:13 crc kubenswrapper[4741]: I0929 20:34:13.570462 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"39cc47ab-05f6-4582-aabc-71217551efc7","Type":"ContainerStarted","Data":"8e9ba2fd429bfabdfb6e3188196d71ee543cc555aa407eb54a29ca91bd57df31"} Sep 29 20:34:13 crc kubenswrapper[4741]: I0929 20:34:13.570635 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dbhjw" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="registry-server" containerID="cri-o://af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f" gracePeriod=2 Sep 29 20:34:13 crc kubenswrapper[4741]: I0929 20:34:13.594347 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.594316134 podStartE2EDuration="2.594316134s" podCreationTimestamp="2025-09-29 20:34:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:13.587759918 +0000 UTC m=+5095.235549290" watchObservedRunningTime="2025-09-29 20:34:13.594316134 +0000 UTC m=+5095.242105506" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.038496 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.199254 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-catalog-content\") pod \"eef7aa30-420d-4e5f-b24b-45401527b15e\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.199454 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnzcr\" (UniqueName: \"kubernetes.io/projected/eef7aa30-420d-4e5f-b24b-45401527b15e-kube-api-access-pnzcr\") pod \"eef7aa30-420d-4e5f-b24b-45401527b15e\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.199539 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-utilities\") pod \"eef7aa30-420d-4e5f-b24b-45401527b15e\" (UID: \"eef7aa30-420d-4e5f-b24b-45401527b15e\") " Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.200225 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-utilities" (OuterVolumeSpecName: "utilities") pod "eef7aa30-420d-4e5f-b24b-45401527b15e" (UID: "eef7aa30-420d-4e5f-b24b-45401527b15e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.207954 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eef7aa30-420d-4e5f-b24b-45401527b15e-kube-api-access-pnzcr" (OuterVolumeSpecName: "kube-api-access-pnzcr") pod "eef7aa30-420d-4e5f-b24b-45401527b15e" (UID: "eef7aa30-420d-4e5f-b24b-45401527b15e"). InnerVolumeSpecName "kube-api-access-pnzcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.257628 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eef7aa30-420d-4e5f-b24b-45401527b15e" (UID: "eef7aa30-420d-4e5f-b24b-45401527b15e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.302133 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.302175 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnzcr\" (UniqueName: \"kubernetes.io/projected/eef7aa30-420d-4e5f-b24b-45401527b15e-kube-api-access-pnzcr\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.302188 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7aa30-420d-4e5f-b24b-45401527b15e-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.579199 4741 generic.go:334] "Generic (PLEG): container finished" podID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerID="af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f" exitCode=0 Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.579251 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbhjw" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.579285 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbhjw" event={"ID":"eef7aa30-420d-4e5f-b24b-45401527b15e","Type":"ContainerDied","Data":"af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f"} Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.579321 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbhjw" event={"ID":"eef7aa30-420d-4e5f-b24b-45401527b15e","Type":"ContainerDied","Data":"8db659a449519b541f190311bf86475754dd27cb69d52f8cdd1f099ec7a2b43f"} Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.579367 4741 scope.go:117] "RemoveContainer" containerID="af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.600662 4741 scope.go:117] "RemoveContainer" containerID="d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.614188 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbhjw"] Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.621714 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dbhjw"] Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.643306 4741 scope.go:117] "RemoveContainer" containerID="f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.658271 4741 scope.go:117] "RemoveContainer" containerID="af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f" Sep 29 20:34:14 crc kubenswrapper[4741]: E0929 20:34:14.658711 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f\": container with ID starting with af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f not found: ID does not exist" containerID="af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.658751 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f"} err="failed to get container status \"af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f\": rpc error: code = NotFound desc = could not find container \"af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f\": container with ID starting with af2c6f8692e23cbcf4b49b9af874849975a8adf9a62c87067526c2d90a551c0f not found: ID does not exist" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.658775 4741 scope.go:117] "RemoveContainer" containerID="d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad" Sep 29 20:34:14 crc kubenswrapper[4741]: E0929 20:34:14.658994 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad\": container with ID starting with d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad not found: ID does not exist" containerID="d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.659023 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad"} err="failed to get container status \"d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad\": rpc error: code = NotFound desc = could not find container \"d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad\": container with ID starting with d2518b834f3b71e634617d3aed79167adbdc4befe0dbe24f7690fca7990611ad not found: ID does not exist" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.659043 4741 scope.go:117] "RemoveContainer" containerID="f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5" Sep 29 20:34:14 crc kubenswrapper[4741]: E0929 20:34:14.659206 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5\": container with ID starting with f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5 not found: ID does not exist" containerID="f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5" Sep 29 20:34:14 crc kubenswrapper[4741]: I0929 20:34:14.659225 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5"} err="failed to get container status \"f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5\": rpc error: code = NotFound desc = could not find container \"f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5\": container with ID starting with f16ab375ab7cd4829fc16fc9dc13aedf76c16bbc337fb7944ac55b25df01c6f5 not found: ID does not exist" Sep 29 20:34:15 crc kubenswrapper[4741]: I0929 20:34:15.096156 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" path="/var/lib/kubelet/pods/eef7aa30-420d-4e5f-b24b-45401527b15e/volumes" Sep 29 20:34:15 crc kubenswrapper[4741]: I0929 20:34:15.544566 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:15 crc kubenswrapper[4741]: I0929 20:34:15.601841 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-5dltm"] Sep 29 20:34:15 crc kubenswrapper[4741]: I0929 20:34:15.603511 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" podUID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerName="dnsmasq-dns" containerID="cri-o://22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b" gracePeriod=10 Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.080674 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.232819 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-config\") pod \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.232996 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-dns-svc\") pod \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.233060 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4h9q\" (UniqueName: \"kubernetes.io/projected/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-kube-api-access-h4h9q\") pod \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\" (UID: \"f2e5500e-0a6f-4ba9-a38e-7339cd30e215\") " Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.238658 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-kube-api-access-h4h9q" (OuterVolumeSpecName: "kube-api-access-h4h9q") pod "f2e5500e-0a6f-4ba9-a38e-7339cd30e215" (UID: "f2e5500e-0a6f-4ba9-a38e-7339cd30e215"). InnerVolumeSpecName "kube-api-access-h4h9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.277126 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-config" (OuterVolumeSpecName: "config") pod "f2e5500e-0a6f-4ba9-a38e-7339cd30e215" (UID: "f2e5500e-0a6f-4ba9-a38e-7339cd30e215"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.281558 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f2e5500e-0a6f-4ba9-a38e-7339cd30e215" (UID: "f2e5500e-0a6f-4ba9-a38e-7339cd30e215"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.334890 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.334939 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4h9q\" (UniqueName: \"kubernetes.io/projected/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-kube-api-access-h4h9q\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.334959 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e5500e-0a6f-4ba9-a38e-7339cd30e215-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.624659 4741 generic.go:334] "Generic (PLEG): container finished" podID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerID="22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b" exitCode=0 Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.624786 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" event={"ID":"f2e5500e-0a6f-4ba9-a38e-7339cd30e215","Type":"ContainerDied","Data":"22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b"} Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.624963 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" event={"ID":"f2e5500e-0a6f-4ba9-a38e-7339cd30e215","Type":"ContainerDied","Data":"db50541e1de20661f73a1482dce47c6afe5385c9c1f6d4a4f32c3ede80a31071"} Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.624986 4741 scope.go:117] "RemoveContainer" containerID="22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.624826 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-5dltm" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.661621 4741 scope.go:117] "RemoveContainer" containerID="02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.682976 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-5dltm"] Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.693972 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-5dltm"] Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.700144 4741 scope.go:117] "RemoveContainer" containerID="22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b" Sep 29 20:34:16 crc kubenswrapper[4741]: E0929 20:34:16.700710 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b\": container with ID starting with 22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b not found: ID does not exist" containerID="22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.700758 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b"} err="failed to get container status \"22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b\": rpc error: code = NotFound desc = could not find container \"22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b\": container with ID starting with 22359ef82dc4642582e89f0cd1053ef24ca67da4960108e2f39f15150488607b not found: ID does not exist" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.700796 4741 scope.go:117] "RemoveContainer" containerID="02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6" Sep 29 20:34:16 crc kubenswrapper[4741]: E0929 20:34:16.701290 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6\": container with ID starting with 02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6 not found: ID does not exist" containerID="02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6" Sep 29 20:34:16 crc kubenswrapper[4741]: I0929 20:34:16.701329 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6"} err="failed to get container status \"02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6\": rpc error: code = NotFound desc = could not find container \"02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6\": container with ID starting with 02019381fcdc3a9d1261518cc279db58abf9e79d8c624f0e9fe6178491b5d9d6 not found: ID does not exist" Sep 29 20:34:17 crc kubenswrapper[4741]: I0929 20:34:17.095915 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" path="/var/lib/kubelet/pods/f2e5500e-0a6f-4ba9-a38e-7339cd30e215/volumes" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521186 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Sep 29 20:34:19 crc kubenswrapper[4741]: E0929 20:34:19.521747 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="extract-content" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521760 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="extract-content" Sep 29 20:34:19 crc kubenswrapper[4741]: E0929 20:34:19.521772 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerName="init" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521778 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerName="init" Sep 29 20:34:19 crc kubenswrapper[4741]: E0929 20:34:19.521809 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="extract-utilities" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521816 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="extract-utilities" Sep 29 20:34:19 crc kubenswrapper[4741]: E0929 20:34:19.521825 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="registry-server" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521831 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="registry-server" Sep 29 20:34:19 crc kubenswrapper[4741]: E0929 20:34:19.521848 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerName="dnsmasq-dns" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521853 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerName="dnsmasq-dns" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521989 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="eef7aa30-420d-4e5f-b24b-45401527b15e" containerName="registry-server" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.521996 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2e5500e-0a6f-4ba9-a38e-7339cd30e215" containerName="dnsmasq-dns" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.523581 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.546046 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5knhm" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.546192 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.549817 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.564498 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.587176 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca6fc6d-897d-4994-abc7-890d66323a76-config\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.587383 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ca6fc6d-897d-4994-abc7-890d66323a76-scripts\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.587575 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ca6fc6d-897d-4994-abc7-890d66323a76-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.587830 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2bxh\" (UniqueName: \"kubernetes.io/projected/4ca6fc6d-897d-4994-abc7-890d66323a76-kube-api-access-t2bxh\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.587977 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6fc6d-897d-4994-abc7-890d66323a76-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.689929 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca6fc6d-897d-4994-abc7-890d66323a76-config\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.689987 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ca6fc6d-897d-4994-abc7-890d66323a76-scripts\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.690036 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ca6fc6d-897d-4994-abc7-890d66323a76-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.690069 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2bxh\" (UniqueName: \"kubernetes.io/projected/4ca6fc6d-897d-4994-abc7-890d66323a76-kube-api-access-t2bxh\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.690093 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6fc6d-897d-4994-abc7-890d66323a76-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.690876 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca6fc6d-897d-4994-abc7-890d66323a76-config\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.690887 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ca6fc6d-897d-4994-abc7-890d66323a76-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.690952 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ca6fc6d-897d-4994-abc7-890d66323a76-scripts\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.696823 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6fc6d-897d-4994-abc7-890d66323a76-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.708558 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2bxh\" (UniqueName: \"kubernetes.io/projected/4ca6fc6d-897d-4994-abc7-890d66323a76-kube-api-access-t2bxh\") pod \"ovn-northd-0\" (UID: \"4ca6fc6d-897d-4994-abc7-890d66323a76\") " pod="openstack/ovn-northd-0" Sep 29 20:34:19 crc kubenswrapper[4741]: I0929 20:34:19.846290 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 20:34:20 crc kubenswrapper[4741]: I0929 20:34:20.293938 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 20:34:20 crc kubenswrapper[4741]: I0929 20:34:20.663755 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4ca6fc6d-897d-4994-abc7-890d66323a76","Type":"ContainerStarted","Data":"1de72b22724e2fc2d1654dbd177cec68bd94e933f4a7dc74c4b1538098163cbf"} Sep 29 20:34:20 crc kubenswrapper[4741]: I0929 20:34:20.664371 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Sep 29 20:34:20 crc kubenswrapper[4741]: I0929 20:34:20.664424 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4ca6fc6d-897d-4994-abc7-890d66323a76","Type":"ContainerStarted","Data":"abc6cd1b3ed40746602446ff1e3790f7212c781833ed70cf3880abf1d6148121"} Sep 29 20:34:20 crc kubenswrapper[4741]: I0929 20:34:20.664439 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4ca6fc6d-897d-4994-abc7-890d66323a76","Type":"ContainerStarted","Data":"49befad53418d68f2bf8897c040e607f6c1460ddd66a424c6446ccaf8c2e262a"} Sep 29 20:34:20 crc kubenswrapper[4741]: I0929 20:34:20.689687 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.689658847 podStartE2EDuration="1.689658847s" podCreationTimestamp="2025-09-29 20:34:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:20.683621297 +0000 UTC m=+5102.331410669" watchObservedRunningTime="2025-09-29 20:34:20.689658847 +0000 UTC m=+5102.337448219" Sep 29 20:34:25 crc kubenswrapper[4741]: I0929 20:34:25.515903 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-pgfbv"] Sep 29 20:34:25 crc kubenswrapper[4741]: I0929 20:34:25.517639 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pgfbv" Sep 29 20:34:25 crc kubenswrapper[4741]: I0929 20:34:25.525892 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-pgfbv"] Sep 29 20:34:25 crc kubenswrapper[4741]: I0929 20:34:25.598250 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c8c8\" (UniqueName: \"kubernetes.io/projected/08f96dea-793e-46e3-9249-033ba38c6607-kube-api-access-4c8c8\") pod \"keystone-db-create-pgfbv\" (UID: \"08f96dea-793e-46e3-9249-033ba38c6607\") " pod="openstack/keystone-db-create-pgfbv" Sep 29 20:34:25 crc kubenswrapper[4741]: I0929 20:34:25.699713 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c8c8\" (UniqueName: \"kubernetes.io/projected/08f96dea-793e-46e3-9249-033ba38c6607-kube-api-access-4c8c8\") pod \"keystone-db-create-pgfbv\" (UID: \"08f96dea-793e-46e3-9249-033ba38c6607\") " pod="openstack/keystone-db-create-pgfbv" Sep 29 20:34:25 crc kubenswrapper[4741]: I0929 20:34:25.719944 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c8c8\" (UniqueName: \"kubernetes.io/projected/08f96dea-793e-46e3-9249-033ba38c6607-kube-api-access-4c8c8\") pod \"keystone-db-create-pgfbv\" (UID: \"08f96dea-793e-46e3-9249-033ba38c6607\") " pod="openstack/keystone-db-create-pgfbv" Sep 29 20:34:25 crc kubenswrapper[4741]: I0929 20:34:25.838217 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pgfbv" Sep 29 20:34:26 crc kubenswrapper[4741]: I0929 20:34:26.262474 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-pgfbv"] Sep 29 20:34:26 crc kubenswrapper[4741]: W0929 20:34:26.266180 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08f96dea_793e_46e3_9249_033ba38c6607.slice/crio-59b51df912fd26a8b94c3e65168ea31da8ae0627f3c2e801d54ff3b15ad9a2ee WatchSource:0}: Error finding container 59b51df912fd26a8b94c3e65168ea31da8ae0627f3c2e801d54ff3b15ad9a2ee: Status 404 returned error can't find the container with id 59b51df912fd26a8b94c3e65168ea31da8ae0627f3c2e801d54ff3b15ad9a2ee Sep 29 20:34:26 crc kubenswrapper[4741]: I0929 20:34:26.715735 4741 generic.go:334] "Generic (PLEG): container finished" podID="08f96dea-793e-46e3-9249-033ba38c6607" containerID="0dc0f530e90466049373310a4126962ad5086cc3a72062fe81b0e165d9ca10a1" exitCode=0 Sep 29 20:34:26 crc kubenswrapper[4741]: I0929 20:34:26.715806 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pgfbv" event={"ID":"08f96dea-793e-46e3-9249-033ba38c6607","Type":"ContainerDied","Data":"0dc0f530e90466049373310a4126962ad5086cc3a72062fe81b0e165d9ca10a1"} Sep 29 20:34:26 crc kubenswrapper[4741]: I0929 20:34:26.715850 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pgfbv" event={"ID":"08f96dea-793e-46e3-9249-033ba38c6607","Type":"ContainerStarted","Data":"59b51df912fd26a8b94c3e65168ea31da8ae0627f3c2e801d54ff3b15ad9a2ee"} Sep 29 20:34:28 crc kubenswrapper[4741]: I0929 20:34:28.080712 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pgfbv" Sep 29 20:34:28 crc kubenswrapper[4741]: I0929 20:34:28.137515 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c8c8\" (UniqueName: \"kubernetes.io/projected/08f96dea-793e-46e3-9249-033ba38c6607-kube-api-access-4c8c8\") pod \"08f96dea-793e-46e3-9249-033ba38c6607\" (UID: \"08f96dea-793e-46e3-9249-033ba38c6607\") " Sep 29 20:34:28 crc kubenswrapper[4741]: I0929 20:34:28.143425 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08f96dea-793e-46e3-9249-033ba38c6607-kube-api-access-4c8c8" (OuterVolumeSpecName: "kube-api-access-4c8c8") pod "08f96dea-793e-46e3-9249-033ba38c6607" (UID: "08f96dea-793e-46e3-9249-033ba38c6607"). InnerVolumeSpecName "kube-api-access-4c8c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:28 crc kubenswrapper[4741]: I0929 20:34:28.240463 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c8c8\" (UniqueName: \"kubernetes.io/projected/08f96dea-793e-46e3-9249-033ba38c6607-kube-api-access-4c8c8\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:28 crc kubenswrapper[4741]: I0929 20:34:28.736857 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pgfbv" event={"ID":"08f96dea-793e-46e3-9249-033ba38c6607","Type":"ContainerDied","Data":"59b51df912fd26a8b94c3e65168ea31da8ae0627f3c2e801d54ff3b15ad9a2ee"} Sep 29 20:34:28 crc kubenswrapper[4741]: I0929 20:34:28.736932 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59b51df912fd26a8b94c3e65168ea31da8ae0627f3c2e801d54ff3b15ad9a2ee" Sep 29 20:34:28 crc kubenswrapper[4741]: I0929 20:34:28.736942 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pgfbv" Sep 29 20:34:34 crc kubenswrapper[4741]: I0929 20:34:34.905077 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.559972 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9b22-account-create-55fn4"] Sep 29 20:34:35 crc kubenswrapper[4741]: E0929 20:34:35.560503 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f96dea-793e-46e3-9249-033ba38c6607" containerName="mariadb-database-create" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.560530 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f96dea-793e-46e3-9249-033ba38c6607" containerName="mariadb-database-create" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.560831 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f96dea-793e-46e3-9249-033ba38c6607" containerName="mariadb-database-create" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.561860 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b22-account-create-55fn4" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.563994 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.573069 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9b22-account-create-55fn4"] Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.599911 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69tbp\" (UniqueName: \"kubernetes.io/projected/afefc204-e529-41d6-9592-22c926ccbf25-kube-api-access-69tbp\") pod \"keystone-9b22-account-create-55fn4\" (UID: \"afefc204-e529-41d6-9592-22c926ccbf25\") " pod="openstack/keystone-9b22-account-create-55fn4" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.700644 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69tbp\" (UniqueName: \"kubernetes.io/projected/afefc204-e529-41d6-9592-22c926ccbf25-kube-api-access-69tbp\") pod \"keystone-9b22-account-create-55fn4\" (UID: \"afefc204-e529-41d6-9592-22c926ccbf25\") " pod="openstack/keystone-9b22-account-create-55fn4" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.721041 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69tbp\" (UniqueName: \"kubernetes.io/projected/afefc204-e529-41d6-9592-22c926ccbf25-kube-api-access-69tbp\") pod \"keystone-9b22-account-create-55fn4\" (UID: \"afefc204-e529-41d6-9592-22c926ccbf25\") " pod="openstack/keystone-9b22-account-create-55fn4" Sep 29 20:34:35 crc kubenswrapper[4741]: I0929 20:34:35.920092 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b22-account-create-55fn4" Sep 29 20:34:36 crc kubenswrapper[4741]: I0929 20:34:36.368799 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9b22-account-create-55fn4"] Sep 29 20:34:36 crc kubenswrapper[4741]: I0929 20:34:36.819120 4741 generic.go:334] "Generic (PLEG): container finished" podID="afefc204-e529-41d6-9592-22c926ccbf25" containerID="a32763cb55fdc7bb481856264c284ca1ec29d9aa5fc4b5d51a06e1b21a9454d3" exitCode=0 Sep 29 20:34:36 crc kubenswrapper[4741]: I0929 20:34:36.819186 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9b22-account-create-55fn4" event={"ID":"afefc204-e529-41d6-9592-22c926ccbf25","Type":"ContainerDied","Data":"a32763cb55fdc7bb481856264c284ca1ec29d9aa5fc4b5d51a06e1b21a9454d3"} Sep 29 20:34:36 crc kubenswrapper[4741]: I0929 20:34:36.819671 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9b22-account-create-55fn4" event={"ID":"afefc204-e529-41d6-9592-22c926ccbf25","Type":"ContainerStarted","Data":"4944933b9ad5c6b74f9469bc5ad334f70b448875457756dc077cddf64efdb93c"} Sep 29 20:34:38 crc kubenswrapper[4741]: I0929 20:34:38.208280 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b22-account-create-55fn4" Sep 29 20:34:38 crc kubenswrapper[4741]: I0929 20:34:38.351427 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69tbp\" (UniqueName: \"kubernetes.io/projected/afefc204-e529-41d6-9592-22c926ccbf25-kube-api-access-69tbp\") pod \"afefc204-e529-41d6-9592-22c926ccbf25\" (UID: \"afefc204-e529-41d6-9592-22c926ccbf25\") " Sep 29 20:34:38 crc kubenswrapper[4741]: I0929 20:34:38.357084 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afefc204-e529-41d6-9592-22c926ccbf25-kube-api-access-69tbp" (OuterVolumeSpecName: "kube-api-access-69tbp") pod "afefc204-e529-41d6-9592-22c926ccbf25" (UID: "afefc204-e529-41d6-9592-22c926ccbf25"). InnerVolumeSpecName "kube-api-access-69tbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:38 crc kubenswrapper[4741]: I0929 20:34:38.453364 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69tbp\" (UniqueName: \"kubernetes.io/projected/afefc204-e529-41d6-9592-22c926ccbf25-kube-api-access-69tbp\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:38 crc kubenswrapper[4741]: I0929 20:34:38.842519 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9b22-account-create-55fn4" event={"ID":"afefc204-e529-41d6-9592-22c926ccbf25","Type":"ContainerDied","Data":"4944933b9ad5c6b74f9469bc5ad334f70b448875457756dc077cddf64efdb93c"} Sep 29 20:34:38 crc kubenswrapper[4741]: I0929 20:34:38.842587 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4944933b9ad5c6b74f9469bc5ad334f70b448875457756dc077cddf64efdb93c" Sep 29 20:34:38 crc kubenswrapper[4741]: I0929 20:34:38.842690 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b22-account-create-55fn4" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.016560 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-qbkn6"] Sep 29 20:34:41 crc kubenswrapper[4741]: E0929 20:34:41.016898 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afefc204-e529-41d6-9592-22c926ccbf25" containerName="mariadb-account-create" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.016910 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="afefc204-e529-41d6-9592-22c926ccbf25" containerName="mariadb-account-create" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.017074 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="afefc204-e529-41d6-9592-22c926ccbf25" containerName="mariadb-account-create" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.017659 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.019871 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ngqvx" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.020053 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.020277 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.020491 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.031429 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qbkn6"] Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.098325 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-combined-ca-bundle\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.098378 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-config-data\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.098441 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6n2n\" (UniqueName: \"kubernetes.io/projected/b16a61e9-d870-4046-8cda-25d5980fb655-kube-api-access-f6n2n\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.199636 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-combined-ca-bundle\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.199703 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-config-data\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.199746 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6n2n\" (UniqueName: \"kubernetes.io/projected/b16a61e9-d870-4046-8cda-25d5980fb655-kube-api-access-f6n2n\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.205418 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-combined-ca-bundle\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.206236 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-config-data\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.216255 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6n2n\" (UniqueName: \"kubernetes.io/projected/b16a61e9-d870-4046-8cda-25d5980fb655-kube-api-access-f6n2n\") pod \"keystone-db-sync-qbkn6\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.337895 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.748945 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qbkn6"] Sep 29 20:34:41 crc kubenswrapper[4741]: I0929 20:34:41.874409 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qbkn6" event={"ID":"b16a61e9-d870-4046-8cda-25d5980fb655","Type":"ContainerStarted","Data":"066523dae334d925787f0a9a586e4ccc168acf42da96e05e2625e467cc3dc6f9"} Sep 29 20:34:42 crc kubenswrapper[4741]: I0929 20:34:42.889864 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qbkn6" event={"ID":"b16a61e9-d870-4046-8cda-25d5980fb655","Type":"ContainerStarted","Data":"64c212e1a8175794b79a0d1c3960b1990d47cdf75ba74c0bdffd833d8fdac26c"} Sep 29 20:34:42 crc kubenswrapper[4741]: I0929 20:34:42.920720 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-qbkn6" podStartSLOduration=2.920691923 podStartE2EDuration="2.920691923s" podCreationTimestamp="2025-09-29 20:34:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:42.912119662 +0000 UTC m=+5124.559909064" watchObservedRunningTime="2025-09-29 20:34:42.920691923 +0000 UTC m=+5124.568481295" Sep 29 20:34:43 crc kubenswrapper[4741]: I0929 20:34:43.902199 4741 generic.go:334] "Generic (PLEG): container finished" podID="b16a61e9-d870-4046-8cda-25d5980fb655" containerID="64c212e1a8175794b79a0d1c3960b1990d47cdf75ba74c0bdffd833d8fdac26c" exitCode=0 Sep 29 20:34:43 crc kubenswrapper[4741]: I0929 20:34:43.902249 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qbkn6" event={"ID":"b16a61e9-d870-4046-8cda-25d5980fb655","Type":"ContainerDied","Data":"64c212e1a8175794b79a0d1c3960b1990d47cdf75ba74c0bdffd833d8fdac26c"} Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.203601 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.272816 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-config-data\") pod \"b16a61e9-d870-4046-8cda-25d5980fb655\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.312677 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-config-data" (OuterVolumeSpecName: "config-data") pod "b16a61e9-d870-4046-8cda-25d5980fb655" (UID: "b16a61e9-d870-4046-8cda-25d5980fb655"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.373969 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-combined-ca-bundle\") pod \"b16a61e9-d870-4046-8cda-25d5980fb655\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.374030 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6n2n\" (UniqueName: \"kubernetes.io/projected/b16a61e9-d870-4046-8cda-25d5980fb655-kube-api-access-f6n2n\") pod \"b16a61e9-d870-4046-8cda-25d5980fb655\" (UID: \"b16a61e9-d870-4046-8cda-25d5980fb655\") " Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.374471 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.376719 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b16a61e9-d870-4046-8cda-25d5980fb655-kube-api-access-f6n2n" (OuterVolumeSpecName: "kube-api-access-f6n2n") pod "b16a61e9-d870-4046-8cda-25d5980fb655" (UID: "b16a61e9-d870-4046-8cda-25d5980fb655"). InnerVolumeSpecName "kube-api-access-f6n2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.397834 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b16a61e9-d870-4046-8cda-25d5980fb655" (UID: "b16a61e9-d870-4046-8cda-25d5980fb655"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.476018 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b16a61e9-d870-4046-8cda-25d5980fb655-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.476057 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6n2n\" (UniqueName: \"kubernetes.io/projected/b16a61e9-d870-4046-8cda-25d5980fb655-kube-api-access-f6n2n\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.922677 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qbkn6" event={"ID":"b16a61e9-d870-4046-8cda-25d5980fb655","Type":"ContainerDied","Data":"066523dae334d925787f0a9a586e4ccc168acf42da96e05e2625e467cc3dc6f9"} Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.922722 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="066523dae334d925787f0a9a586e4ccc168acf42da96e05e2625e467cc3dc6f9" Sep 29 20:34:45 crc kubenswrapper[4741]: I0929 20:34:45.922729 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qbkn6" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.133591 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2b9nz"] Sep 29 20:34:46 crc kubenswrapper[4741]: E0929 20:34:46.134198 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b16a61e9-d870-4046-8cda-25d5980fb655" containerName="keystone-db-sync" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.134210 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b16a61e9-d870-4046-8cda-25d5980fb655" containerName="keystone-db-sync" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.134375 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b16a61e9-d870-4046-8cda-25d5980fb655" containerName="keystone-db-sync" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.134959 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.138622 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.139460 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.140173 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ngqvx" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.140758 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.158129 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2b9nz"] Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.178327 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bd4f7d747-jx7jl"] Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.180586 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.185647 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdj8l\" (UniqueName: \"kubernetes.io/projected/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-kube-api-access-kdj8l\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.185721 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-nb\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.185766 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-dns-svc\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.185792 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-config\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.185812 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-sb\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.206014 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd4f7d747-jx7jl"] Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287410 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-config-data\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287466 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdj8l\" (UniqueName: \"kubernetes.io/projected/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-kube-api-access-kdj8l\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287508 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-scripts\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287532 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-nb\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287571 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-dns-svc\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287594 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-config\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287618 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-combined-ca-bundle\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287647 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-fernet-keys\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287668 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-sb\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287695 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-credential-keys\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.287721 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwpkb\" (UniqueName: \"kubernetes.io/projected/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-kube-api-access-pwpkb\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.289374 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-nb\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.289889 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-dns-svc\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.290381 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-config\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.291073 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-sb\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.309985 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdj8l\" (UniqueName: \"kubernetes.io/projected/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-kube-api-access-kdj8l\") pod \"dnsmasq-dns-bd4f7d747-jx7jl\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.390472 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-combined-ca-bundle\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.390508 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-fernet-keys\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.390533 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-credential-keys\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.390556 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwpkb\" (UniqueName: \"kubernetes.io/projected/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-kube-api-access-pwpkb\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.390596 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-config-data\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.390633 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-scripts\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.393936 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-scripts\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.394160 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-credential-keys\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.394478 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-combined-ca-bundle\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.394635 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-config-data\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.395265 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-fernet-keys\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.409318 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwpkb\" (UniqueName: \"kubernetes.io/projected/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-kube-api-access-pwpkb\") pod \"keystone-bootstrap-2b9nz\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.506729 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.513092 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:46 crc kubenswrapper[4741]: I0929 20:34:46.988453 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2b9nz"] Sep 29 20:34:47 crc kubenswrapper[4741]: I0929 20:34:47.022971 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd4f7d747-jx7jl"] Sep 29 20:34:47 crc kubenswrapper[4741]: W0929 20:34:47.025439 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod174dfc5c_a80c_4ee1_84c2_427761b0b8ec.slice/crio-7ad444886e518c74fcbb73f6e329056690199d3f8640ee54c66be93f708758e0 WatchSource:0}: Error finding container 7ad444886e518c74fcbb73f6e329056690199d3f8640ee54c66be93f708758e0: Status 404 returned error can't find the container with id 7ad444886e518c74fcbb73f6e329056690199d3f8640ee54c66be93f708758e0 Sep 29 20:34:47 crc kubenswrapper[4741]: I0929 20:34:47.936759 4741 generic.go:334] "Generic (PLEG): container finished" podID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerID="74709b5967068c4804bc8bb6b75ca331018c117aa1f50c70e70390a1cffdbd30" exitCode=0 Sep 29 20:34:47 crc kubenswrapper[4741]: I0929 20:34:47.936918 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" event={"ID":"174dfc5c-a80c-4ee1-84c2-427761b0b8ec","Type":"ContainerDied","Data":"74709b5967068c4804bc8bb6b75ca331018c117aa1f50c70e70390a1cffdbd30"} Sep 29 20:34:47 crc kubenswrapper[4741]: I0929 20:34:47.937029 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" event={"ID":"174dfc5c-a80c-4ee1-84c2-427761b0b8ec","Type":"ContainerStarted","Data":"7ad444886e518c74fcbb73f6e329056690199d3f8640ee54c66be93f708758e0"} Sep 29 20:34:47 crc kubenswrapper[4741]: I0929 20:34:47.938939 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b9nz" event={"ID":"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b","Type":"ContainerStarted","Data":"87771f9128e30a94e6a5ed3e5a856844569a501687ca316d95f545805fa3b759"} Sep 29 20:34:47 crc kubenswrapper[4741]: I0929 20:34:47.938967 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b9nz" event={"ID":"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b","Type":"ContainerStarted","Data":"4c57355d17bb8845e7ddddb1a0def5c962c02afdb4e12c87a6e4e7df0d439e15"} Sep 29 20:34:47 crc kubenswrapper[4741]: I0929 20:34:47.991946 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2b9nz" podStartSLOduration=1.991918204 podStartE2EDuration="1.991918204s" podCreationTimestamp="2025-09-29 20:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:47.975217967 +0000 UTC m=+5129.623007309" watchObservedRunningTime="2025-09-29 20:34:47.991918204 +0000 UTC m=+5129.639707536" Sep 29 20:34:48 crc kubenswrapper[4741]: I0929 20:34:48.949656 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" event={"ID":"174dfc5c-a80c-4ee1-84c2-427761b0b8ec","Type":"ContainerStarted","Data":"7cc7221bda1769b4d8634c9ddfa76c534c09e093f8ee3fbfa9abfa530f1a7325"} Sep 29 20:34:48 crc kubenswrapper[4741]: I0929 20:34:48.951438 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:48 crc kubenswrapper[4741]: I0929 20:34:48.986366 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" podStartSLOduration=2.986340409 podStartE2EDuration="2.986340409s" podCreationTimestamp="2025-09-29 20:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:48.972811682 +0000 UTC m=+5130.620601054" watchObservedRunningTime="2025-09-29 20:34:48.986340409 +0000 UTC m=+5130.634129781" Sep 29 20:34:50 crc kubenswrapper[4741]: I0929 20:34:50.969077 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b9nz" event={"ID":"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b","Type":"ContainerDied","Data":"87771f9128e30a94e6a5ed3e5a856844569a501687ca316d95f545805fa3b759"} Sep 29 20:34:50 crc kubenswrapper[4741]: I0929 20:34:50.969123 4741 generic.go:334] "Generic (PLEG): container finished" podID="2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" containerID="87771f9128e30a94e6a5ed3e5a856844569a501687ca316d95f545805fa3b759" exitCode=0 Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.347758 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.496144 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-fernet-keys\") pod \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.496275 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-scripts\") pod \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.496318 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-credential-keys\") pod \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.496372 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-combined-ca-bundle\") pod \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.496425 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwpkb\" (UniqueName: \"kubernetes.io/projected/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-kube-api-access-pwpkb\") pod \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.496441 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-config-data\") pod \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\" (UID: \"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b\") " Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.501678 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-scripts" (OuterVolumeSpecName: "scripts") pod "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" (UID: "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.502083 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" (UID: "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.502658 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-kube-api-access-pwpkb" (OuterVolumeSpecName: "kube-api-access-pwpkb") pod "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" (UID: "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b"). InnerVolumeSpecName "kube-api-access-pwpkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.503247 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" (UID: "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.520984 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" (UID: "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.527137 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-config-data" (OuterVolumeSpecName: "config-data") pod "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" (UID: "2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.597881 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.597916 4741 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.597927 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.597937 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwpkb\" (UniqueName: \"kubernetes.io/projected/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-kube-api-access-pwpkb\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.597947 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.597956 4741 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.990986 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b9nz" event={"ID":"2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b","Type":"ContainerDied","Data":"4c57355d17bb8845e7ddddb1a0def5c962c02afdb4e12c87a6e4e7df0d439e15"} Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.991024 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c57355d17bb8845e7ddddb1a0def5c962c02afdb4e12c87a6e4e7df0d439e15" Sep 29 20:34:52 crc kubenswrapper[4741]: I0929 20:34:52.991490 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b9nz" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.074048 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2b9nz"] Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.080596 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2b9nz"] Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.096002 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" path="/var/lib/kubelet/pods/2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b/volumes" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.153869 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nkcj5"] Sep 29 20:34:53 crc kubenswrapper[4741]: E0929 20:34:53.154525 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" containerName="keystone-bootstrap" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.154551 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" containerName="keystone-bootstrap" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.154922 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fbe162d-51ff-4ef8-9a76-a2de45d7aa9b" containerName="keystone-bootstrap" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.155647 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.158161 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ngqvx" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.158309 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.159829 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.160196 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.201725 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nkcj5"] Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.309805 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdfwz\" (UniqueName: \"kubernetes.io/projected/a8650376-c59a-47cd-9ff6-a688a16516ed-kube-api-access-wdfwz\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.310074 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-credential-keys\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.310182 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-config-data\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.310260 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-scripts\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.310352 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-fernet-keys\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.310464 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-combined-ca-bundle\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.411963 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdfwz\" (UniqueName: \"kubernetes.io/projected/a8650376-c59a-47cd-9ff6-a688a16516ed-kube-api-access-wdfwz\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.412010 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-credential-keys\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.412048 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-config-data\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.412075 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-scripts\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.412117 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-fernet-keys\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.412138 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-combined-ca-bundle\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.417018 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-fernet-keys\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.423936 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-credential-keys\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.423947 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-scripts\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.424172 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-combined-ca-bundle\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.427821 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-config-data\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.428261 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdfwz\" (UniqueName: \"kubernetes.io/projected/a8650376-c59a-47cd-9ff6-a688a16516ed-kube-api-access-wdfwz\") pod \"keystone-bootstrap-nkcj5\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.514687 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:53 crc kubenswrapper[4741]: I0929 20:34:53.947254 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nkcj5"] Sep 29 20:34:54 crc kubenswrapper[4741]: I0929 20:34:54.002375 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkcj5" event={"ID":"a8650376-c59a-47cd-9ff6-a688a16516ed","Type":"ContainerStarted","Data":"b9cbd5d94af8593bb9744ea7aebda81b4af19480bdb718a03c9eb1968b5d200e"} Sep 29 20:34:55 crc kubenswrapper[4741]: I0929 20:34:55.013977 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkcj5" event={"ID":"a8650376-c59a-47cd-9ff6-a688a16516ed","Type":"ContainerStarted","Data":"3430ec746ab2a3c23749a224cdb681191d6b55f8b86d5c8b765209f08709bde6"} Sep 29 20:34:55 crc kubenswrapper[4741]: I0929 20:34:55.031628 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nkcj5" podStartSLOduration=2.031603161 podStartE2EDuration="2.031603161s" podCreationTimestamp="2025-09-29 20:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:34:55.030889888 +0000 UTC m=+5136.678679220" watchObservedRunningTime="2025-09-29 20:34:55.031603161 +0000 UTC m=+5136.679392503" Sep 29 20:34:56 crc kubenswrapper[4741]: I0929 20:34:56.515673 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:34:56 crc kubenswrapper[4741]: I0929 20:34:56.582056 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c6dcccb85-lhhw9"] Sep 29 20:34:56 crc kubenswrapper[4741]: I0929 20:34:56.582563 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" podUID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerName="dnsmasq-dns" containerID="cri-o://06a4eefef3fe2023e87f9c4c2cf1a81bf4cfc05e5ea0cae8f9cec2a5659d5bd5" gracePeriod=10 Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.042916 4741 generic.go:334] "Generic (PLEG): container finished" podID="a8650376-c59a-47cd-9ff6-a688a16516ed" containerID="3430ec746ab2a3c23749a224cdb681191d6b55f8b86d5c8b765209f08709bde6" exitCode=0 Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.042993 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkcj5" event={"ID":"a8650376-c59a-47cd-9ff6-a688a16516ed","Type":"ContainerDied","Data":"3430ec746ab2a3c23749a224cdb681191d6b55f8b86d5c8b765209f08709bde6"} Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.045279 4741 generic.go:334] "Generic (PLEG): container finished" podID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerID="06a4eefef3fe2023e87f9c4c2cf1a81bf4cfc05e5ea0cae8f9cec2a5659d5bd5" exitCode=0 Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.045331 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" event={"ID":"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef","Type":"ContainerDied","Data":"06a4eefef3fe2023e87f9c4c2cf1a81bf4cfc05e5ea0cae8f9cec2a5659d5bd5"} Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.108799 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.288745 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-nb\") pod \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.289086 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-sb\") pod \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.289114 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-config\") pod \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.289279 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgzkc\" (UniqueName: \"kubernetes.io/projected/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-kube-api-access-lgzkc\") pod \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.289321 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc\") pod \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.294328 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-kube-api-access-lgzkc" (OuterVolumeSpecName: "kube-api-access-lgzkc") pod "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" (UID: "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef"). InnerVolumeSpecName "kube-api-access-lgzkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.331287 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" (UID: "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.331515 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-config" (OuterVolumeSpecName: "config") pod "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" (UID: "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:57 crc kubenswrapper[4741]: E0929 20:34:57.350026 4741 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc podName:8d4efa1d-6e38-426b-ba9a-8972f3cbbeef nodeName:}" failed. No retries permitted until 2025-09-29 20:34:57.850000243 +0000 UTC m=+5139.497789575 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc") pod "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" (UID: "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef") : error deleting /var/lib/kubelet/pods/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef/volume-subpaths: remove /var/lib/kubelet/pods/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef/volume-subpaths: no such file or directory Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.351136 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" (UID: "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.391420 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.391551 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.391612 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.391685 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgzkc\" (UniqueName: \"kubernetes.io/projected/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-kube-api-access-lgzkc\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.898342 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc\") pod \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\" (UID: \"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef\") " Sep 29 20:34:57 crc kubenswrapper[4741]: I0929 20:34:57.898947 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" (UID: "8d4efa1d-6e38-426b-ba9a-8972f3cbbeef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.000727 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.055923 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.055916 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6dcccb85-lhhw9" event={"ID":"8d4efa1d-6e38-426b-ba9a-8972f3cbbeef","Type":"ContainerDied","Data":"84551570b7eeed6eb3c7e52a180c4a846f4d66e070f951686c4454212d462151"} Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.056038 4741 scope.go:117] "RemoveContainer" containerID="06a4eefef3fe2023e87f9c4c2cf1a81bf4cfc05e5ea0cae8f9cec2a5659d5bd5" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.091667 4741 scope.go:117] "RemoveContainer" containerID="88962a05cee4ae28aa06f8f04b337dff9b4d1501c92178eb5ef81580e028b57a" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.109275 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c6dcccb85-lhhw9"] Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.116590 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c6dcccb85-lhhw9"] Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.371179 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.508521 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-scripts\") pod \"a8650376-c59a-47cd-9ff6-a688a16516ed\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.508599 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-combined-ca-bundle\") pod \"a8650376-c59a-47cd-9ff6-a688a16516ed\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.508651 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdfwz\" (UniqueName: \"kubernetes.io/projected/a8650376-c59a-47cd-9ff6-a688a16516ed-kube-api-access-wdfwz\") pod \"a8650376-c59a-47cd-9ff6-a688a16516ed\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.508679 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-config-data\") pod \"a8650376-c59a-47cd-9ff6-a688a16516ed\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.508709 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-credential-keys\") pod \"a8650376-c59a-47cd-9ff6-a688a16516ed\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.508765 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-fernet-keys\") pod \"a8650376-c59a-47cd-9ff6-a688a16516ed\" (UID: \"a8650376-c59a-47cd-9ff6-a688a16516ed\") " Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.512460 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a8650376-c59a-47cd-9ff6-a688a16516ed" (UID: "a8650376-c59a-47cd-9ff6-a688a16516ed"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.513715 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a8650376-c59a-47cd-9ff6-a688a16516ed" (UID: "a8650376-c59a-47cd-9ff6-a688a16516ed"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.513785 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8650376-c59a-47cd-9ff6-a688a16516ed-kube-api-access-wdfwz" (OuterVolumeSpecName: "kube-api-access-wdfwz") pod "a8650376-c59a-47cd-9ff6-a688a16516ed" (UID: "a8650376-c59a-47cd-9ff6-a688a16516ed"). InnerVolumeSpecName "kube-api-access-wdfwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.515504 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-scripts" (OuterVolumeSpecName: "scripts") pod "a8650376-c59a-47cd-9ff6-a688a16516ed" (UID: "a8650376-c59a-47cd-9ff6-a688a16516ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.532618 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8650376-c59a-47cd-9ff6-a688a16516ed" (UID: "a8650376-c59a-47cd-9ff6-a688a16516ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.534127 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-config-data" (OuterVolumeSpecName: "config-data") pod "a8650376-c59a-47cd-9ff6-a688a16516ed" (UID: "a8650376-c59a-47cd-9ff6-a688a16516ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.610719 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.610778 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.610797 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdfwz\" (UniqueName: \"kubernetes.io/projected/a8650376-c59a-47cd-9ff6-a688a16516ed-kube-api-access-wdfwz\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.610814 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.610831 4741 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:58 crc kubenswrapper[4741]: I0929 20:34:58.610852 4741 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8650376-c59a-47cd-9ff6-a688a16516ed-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.067301 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkcj5" event={"ID":"a8650376-c59a-47cd-9ff6-a688a16516ed","Type":"ContainerDied","Data":"b9cbd5d94af8593bb9744ea7aebda81b4af19480bdb718a03c9eb1968b5d200e"} Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.067633 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9cbd5d94af8593bb9744ea7aebda81b4af19480bdb718a03c9eb1968b5d200e" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.067423 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkcj5" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.095292 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" path="/var/lib/kubelet/pods/8d4efa1d-6e38-426b-ba9a-8972f3cbbeef/volumes" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.169882 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c569dd67b-5gb8r"] Sep 29 20:34:59 crc kubenswrapper[4741]: E0929 20:34:59.170204 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerName="dnsmasq-dns" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.170218 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerName="dnsmasq-dns" Sep 29 20:34:59 crc kubenswrapper[4741]: E0929 20:34:59.170235 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8650376-c59a-47cd-9ff6-a688a16516ed" containerName="keystone-bootstrap" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.170243 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8650376-c59a-47cd-9ff6-a688a16516ed" containerName="keystone-bootstrap" Sep 29 20:34:59 crc kubenswrapper[4741]: E0929 20:34:59.170268 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerName="init" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.170274 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerName="init" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.170447 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d4efa1d-6e38-426b-ba9a-8972f3cbbeef" containerName="dnsmasq-dns" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.170464 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8650376-c59a-47cd-9ff6-a688a16516ed" containerName="keystone-bootstrap" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.171033 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.175863 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.175907 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.179609 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.179778 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ngqvx" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.193469 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c569dd67b-5gb8r"] Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.322871 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-scripts\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.322942 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42xwh\" (UniqueName: \"kubernetes.io/projected/6b9e24cb-db42-4298-ada9-940790341bcd-kube-api-access-42xwh\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.323218 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-fernet-keys\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.323346 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-config-data\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.323465 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-credential-keys\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.323549 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-combined-ca-bundle\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.424999 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-fernet-keys\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.425072 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-config-data\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.426069 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-credential-keys\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.426122 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-combined-ca-bundle\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.426175 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-scripts\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.426275 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42xwh\" (UniqueName: \"kubernetes.io/projected/6b9e24cb-db42-4298-ada9-940790341bcd-kube-api-access-42xwh\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.430168 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-scripts\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.430203 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-fernet-keys\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.430522 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-combined-ca-bundle\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.436005 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-config-data\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.437854 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b9e24cb-db42-4298-ada9-940790341bcd-credential-keys\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.452891 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42xwh\" (UniqueName: \"kubernetes.io/projected/6b9e24cb-db42-4298-ada9-940790341bcd-kube-api-access-42xwh\") pod \"keystone-c569dd67b-5gb8r\" (UID: \"6b9e24cb-db42-4298-ada9-940790341bcd\") " pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.526175 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:34:59 crc kubenswrapper[4741]: I0929 20:34:59.984891 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c569dd67b-5gb8r"] Sep 29 20:35:00 crc kubenswrapper[4741]: I0929 20:35:00.076212 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c569dd67b-5gb8r" event={"ID":"6b9e24cb-db42-4298-ada9-940790341bcd","Type":"ContainerStarted","Data":"c5aa708a3b5ca76cdc79c8d481e0e9f007818ba66986c7f8c1caf8f1ebaa9442"} Sep 29 20:35:01 crc kubenswrapper[4741]: I0929 20:35:01.120353 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c569dd67b-5gb8r" event={"ID":"6b9e24cb-db42-4298-ada9-940790341bcd","Type":"ContainerStarted","Data":"9b246ccbd963562ce173a84312b9a145c5a1460ffda0265ae2e4260607e95e97"} Sep 29 20:35:01 crc kubenswrapper[4741]: I0929 20:35:01.120439 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:35:01 crc kubenswrapper[4741]: I0929 20:35:01.143814 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c569dd67b-5gb8r" podStartSLOduration=2.143793245 podStartE2EDuration="2.143793245s" podCreationTimestamp="2025-09-29 20:34:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:35:01.141042618 +0000 UTC m=+5142.788831950" watchObservedRunningTime="2025-09-29 20:35:01.143793245 +0000 UTC m=+5142.791582577" Sep 29 20:35:26 crc kubenswrapper[4741]: I0929 20:35:26.551485 4741 scope.go:117] "RemoveContainer" containerID="12501422c754abfa4c4bc19b10a7120f658748f1a3abcbf9334cf7321e23e133" Sep 29 20:35:26 crc kubenswrapper[4741]: I0929 20:35:26.586982 4741 scope.go:117] "RemoveContainer" containerID="9dbca03c3d04d4330f9bb1ce6a2d03ad08690e23e6f52a9afa980302611c11bc" Sep 29 20:35:26 crc kubenswrapper[4741]: I0929 20:35:26.628347 4741 scope.go:117] "RemoveContainer" containerID="2856bb2c0447bc7c79e50745235a9a547ffd3f7bc9456d3ae05bd43e3dd3b98c" Sep 29 20:35:26 crc kubenswrapper[4741]: I0929 20:35:26.661664 4741 scope.go:117] "RemoveContainer" containerID="3a7a00fdb32edde507bc32b885dd92dec63a439b9f4991531215115bf39a7bce" Sep 29 20:35:26 crc kubenswrapper[4741]: I0929 20:35:26.692354 4741 scope.go:117] "RemoveContainer" containerID="ab3382abf320ff83b6d2751c76f189021d8921fca2886218c2d19a6dcb670fb3" Sep 29 20:35:31 crc kubenswrapper[4741]: I0929 20:35:31.031557 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-c569dd67b-5gb8r" Sep 29 20:35:31 crc kubenswrapper[4741]: I0929 20:35:31.739620 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:35:31 crc kubenswrapper[4741]: I0929 20:35:31.739711 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.580715 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.582833 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.585363 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.586190 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.586215 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-4z6gc" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.590017 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.678093 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config-secret\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.678180 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.678284 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtmp\" (UniqueName: \"kubernetes.io/projected/65ecdd59-688a-45c4-bd78-a4a0a27f338c-kube-api-access-qgtmp\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.779518 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config-secret\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.779595 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.779653 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgtmp\" (UniqueName: \"kubernetes.io/projected/65ecdd59-688a-45c4-bd78-a4a0a27f338c-kube-api-access-qgtmp\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.780751 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.786488 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config-secret\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.798679 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgtmp\" (UniqueName: \"kubernetes.io/projected/65ecdd59-688a-45c4-bd78-a4a0a27f338c-kube-api-access-qgtmp\") pod \"openstackclient\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.931658 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.945162 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j86v9"] Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.970552 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:35 crc kubenswrapper[4741]: I0929 20:35:35.974454 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j86v9"] Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.085594 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-utilities\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.085648 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-catalog-content\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.085689 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxrph\" (UniqueName: \"kubernetes.io/projected/732f829a-1021-47d7-b9b7-266405d9557f-kube-api-access-nxrph\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.187222 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-catalog-content\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.187304 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxrph\" (UniqueName: \"kubernetes.io/projected/732f829a-1021-47d7-b9b7-266405d9557f-kube-api-access-nxrph\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.187443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-utilities\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.187843 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-utilities\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.188401 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-catalog-content\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.204835 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxrph\" (UniqueName: \"kubernetes.io/projected/732f829a-1021-47d7-b9b7-266405d9557f-kube-api-access-nxrph\") pod \"redhat-marketplace-j86v9\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.334923 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-btkdc"] Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.336489 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.352831 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-btkdc"] Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.386311 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.413175 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.455657 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"65ecdd59-688a-45c4-bd78-a4a0a27f338c","Type":"ContainerStarted","Data":"01134779c188bca21bf592efac8d20f972a182d5b2a6ede95032572ab76091b5"} Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.513371 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qgt2\" (UniqueName: \"kubernetes.io/projected/20269d23-3f5c-41bb-ab7f-34eb83010a73-kube-api-access-7qgt2\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.513446 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-utilities\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.513470 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-catalog-content\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.638282 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qgt2\" (UniqueName: \"kubernetes.io/projected/20269d23-3f5c-41bb-ab7f-34eb83010a73-kube-api-access-7qgt2\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.639343 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-utilities\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.638370 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-utilities\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.639664 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-catalog-content\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.640010 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-catalog-content\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.660272 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qgt2\" (UniqueName: \"kubernetes.io/projected/20269d23-3f5c-41bb-ab7f-34eb83010a73-kube-api-access-7qgt2\") pod \"redhat-operators-btkdc\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.870063 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j86v9"] Sep 29 20:35:36 crc kubenswrapper[4741]: I0929 20:35:36.956689 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:37 crc kubenswrapper[4741]: I0929 20:35:37.467807 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"65ecdd59-688a-45c4-bd78-a4a0a27f338c","Type":"ContainerStarted","Data":"eaef049f3ddd3bc029b384e13683dd571330298cf39a43da1c0596d11b82e9bb"} Sep 29 20:35:37 crc kubenswrapper[4741]: I0929 20:35:37.469279 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-btkdc"] Sep 29 20:35:37 crc kubenswrapper[4741]: I0929 20:35:37.474121 4741 generic.go:334] "Generic (PLEG): container finished" podID="732f829a-1021-47d7-b9b7-266405d9557f" containerID="592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676" exitCode=0 Sep 29 20:35:37 crc kubenswrapper[4741]: I0929 20:35:37.474182 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j86v9" event={"ID":"732f829a-1021-47d7-b9b7-266405d9557f","Type":"ContainerDied","Data":"592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676"} Sep 29 20:35:37 crc kubenswrapper[4741]: I0929 20:35:37.474213 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j86v9" event={"ID":"732f829a-1021-47d7-b9b7-266405d9557f","Type":"ContainerStarted","Data":"d229224979c69432903b306bf9b631007c5feda36014d0ee41be8fdc417944c2"} Sep 29 20:35:37 crc kubenswrapper[4741]: I0929 20:35:37.475792 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:35:37 crc kubenswrapper[4741]: I0929 20:35:37.492231 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.492215737 podStartE2EDuration="2.492215737s" podCreationTimestamp="2025-09-29 20:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:35:37.488926984 +0000 UTC m=+5179.136716316" watchObservedRunningTime="2025-09-29 20:35:37.492215737 +0000 UTC m=+5179.140005069" Sep 29 20:35:38 crc kubenswrapper[4741]: I0929 20:35:38.484020 4741 generic.go:334] "Generic (PLEG): container finished" podID="732f829a-1021-47d7-b9b7-266405d9557f" containerID="ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50" exitCode=0 Sep 29 20:35:38 crc kubenswrapper[4741]: I0929 20:35:38.484177 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j86v9" event={"ID":"732f829a-1021-47d7-b9b7-266405d9557f","Type":"ContainerDied","Data":"ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50"} Sep 29 20:35:38 crc kubenswrapper[4741]: I0929 20:35:38.486617 4741 generic.go:334] "Generic (PLEG): container finished" podID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerID="3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc" exitCode=0 Sep 29 20:35:38 crc kubenswrapper[4741]: I0929 20:35:38.486734 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btkdc" event={"ID":"20269d23-3f5c-41bb-ab7f-34eb83010a73","Type":"ContainerDied","Data":"3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc"} Sep 29 20:35:38 crc kubenswrapper[4741]: I0929 20:35:38.486782 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btkdc" event={"ID":"20269d23-3f5c-41bb-ab7f-34eb83010a73","Type":"ContainerStarted","Data":"62fe0c14b8336e0ffc3c4296b20c65f8f318813c9afbcc0417af8607cfd69c48"} Sep 29 20:35:39 crc kubenswrapper[4741]: I0929 20:35:39.495042 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btkdc" event={"ID":"20269d23-3f5c-41bb-ab7f-34eb83010a73","Type":"ContainerStarted","Data":"b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad"} Sep 29 20:35:39 crc kubenswrapper[4741]: I0929 20:35:39.497233 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j86v9" event={"ID":"732f829a-1021-47d7-b9b7-266405d9557f","Type":"ContainerStarted","Data":"d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44"} Sep 29 20:35:39 crc kubenswrapper[4741]: I0929 20:35:39.536906 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j86v9" podStartSLOduration=2.985696547 podStartE2EDuration="4.536887968s" podCreationTimestamp="2025-09-29 20:35:35 +0000 UTC" firstStartedPulling="2025-09-29 20:35:37.475536531 +0000 UTC m=+5179.123325863" lastFinishedPulling="2025-09-29 20:35:39.026727952 +0000 UTC m=+5180.674517284" observedRunningTime="2025-09-29 20:35:39.530862528 +0000 UTC m=+5181.178651870" watchObservedRunningTime="2025-09-29 20:35:39.536887968 +0000 UTC m=+5181.184677300" Sep 29 20:35:40 crc kubenswrapper[4741]: I0929 20:35:40.506905 4741 generic.go:334] "Generic (PLEG): container finished" podID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerID="b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad" exitCode=0 Sep 29 20:35:40 crc kubenswrapper[4741]: I0929 20:35:40.507001 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btkdc" event={"ID":"20269d23-3f5c-41bb-ab7f-34eb83010a73","Type":"ContainerDied","Data":"b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad"} Sep 29 20:35:41 crc kubenswrapper[4741]: I0929 20:35:41.516478 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btkdc" event={"ID":"20269d23-3f5c-41bb-ab7f-34eb83010a73","Type":"ContainerStarted","Data":"1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557"} Sep 29 20:35:46 crc kubenswrapper[4741]: I0929 20:35:46.387338 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:46 crc kubenswrapper[4741]: I0929 20:35:46.387970 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:46 crc kubenswrapper[4741]: I0929 20:35:46.428062 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:46 crc kubenswrapper[4741]: I0929 20:35:46.445081 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-btkdc" podStartSLOduration=7.9690803599999995 podStartE2EDuration="10.44506013s" podCreationTimestamp="2025-09-29 20:35:36 +0000 UTC" firstStartedPulling="2025-09-29 20:35:38.491225657 +0000 UTC m=+5180.139014989" lastFinishedPulling="2025-09-29 20:35:40.967205417 +0000 UTC m=+5182.614994759" observedRunningTime="2025-09-29 20:35:41.548702492 +0000 UTC m=+5183.196491814" watchObservedRunningTime="2025-09-29 20:35:46.44506013 +0000 UTC m=+5188.092849482" Sep 29 20:35:46 crc kubenswrapper[4741]: I0929 20:35:46.637611 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:46 crc kubenswrapper[4741]: I0929 20:35:46.957728 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:46 crc kubenswrapper[4741]: I0929 20:35:46.957958 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:47 crc kubenswrapper[4741]: I0929 20:35:47.002172 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:47 crc kubenswrapper[4741]: I0929 20:35:47.626056 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:49 crc kubenswrapper[4741]: I0929 20:35:49.329303 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-btkdc"] Sep 29 20:35:49 crc kubenswrapper[4741]: I0929 20:35:49.927135 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j86v9"] Sep 29 20:35:49 crc kubenswrapper[4741]: I0929 20:35:49.927475 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j86v9" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="registry-server" containerID="cri-o://d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44" gracePeriod=2 Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.364662 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.519608 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-utilities\") pod \"732f829a-1021-47d7-b9b7-266405d9557f\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.519713 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxrph\" (UniqueName: \"kubernetes.io/projected/732f829a-1021-47d7-b9b7-266405d9557f-kube-api-access-nxrph\") pod \"732f829a-1021-47d7-b9b7-266405d9557f\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.519840 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-catalog-content\") pod \"732f829a-1021-47d7-b9b7-266405d9557f\" (UID: \"732f829a-1021-47d7-b9b7-266405d9557f\") " Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.521232 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-utilities" (OuterVolumeSpecName: "utilities") pod "732f829a-1021-47d7-b9b7-266405d9557f" (UID: "732f829a-1021-47d7-b9b7-266405d9557f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.528839 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/732f829a-1021-47d7-b9b7-266405d9557f-kube-api-access-nxrph" (OuterVolumeSpecName: "kube-api-access-nxrph") pod "732f829a-1021-47d7-b9b7-266405d9557f" (UID: "732f829a-1021-47d7-b9b7-266405d9557f"). InnerVolumeSpecName "kube-api-access-nxrph". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.542551 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "732f829a-1021-47d7-b9b7-266405d9557f" (UID: "732f829a-1021-47d7-b9b7-266405d9557f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.597441 4741 generic.go:334] "Generic (PLEG): container finished" podID="732f829a-1021-47d7-b9b7-266405d9557f" containerID="d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44" exitCode=0 Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.597526 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j86v9" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.597575 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j86v9" event={"ID":"732f829a-1021-47d7-b9b7-266405d9557f","Type":"ContainerDied","Data":"d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44"} Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.597655 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j86v9" event={"ID":"732f829a-1021-47d7-b9b7-266405d9557f","Type":"ContainerDied","Data":"d229224979c69432903b306bf9b631007c5feda36014d0ee41be8fdc417944c2"} Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.597689 4741 scope.go:117] "RemoveContainer" containerID="d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.597847 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-btkdc" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="registry-server" containerID="cri-o://1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557" gracePeriod=2 Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.622754 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.622799 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxrph\" (UniqueName: \"kubernetes.io/projected/732f829a-1021-47d7-b9b7-266405d9557f-kube-api-access-nxrph\") on node \"crc\" DevicePath \"\"" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.622818 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732f829a-1021-47d7-b9b7-266405d9557f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.635909 4741 scope.go:117] "RemoveContainer" containerID="ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.645761 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j86v9"] Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.656889 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j86v9"] Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.670764 4741 scope.go:117] "RemoveContainer" containerID="592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.704720 4741 scope.go:117] "RemoveContainer" containerID="d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44" Sep 29 20:35:50 crc kubenswrapper[4741]: E0929 20:35:50.705198 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44\": container with ID starting with d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44 not found: ID does not exist" containerID="d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.705237 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44"} err="failed to get container status \"d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44\": rpc error: code = NotFound desc = could not find container \"d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44\": container with ID starting with d9d9873365d1d19a87e37ff7adc7310591f3eb5012538b3d2f8f8f68510b9b44 not found: ID does not exist" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.705264 4741 scope.go:117] "RemoveContainer" containerID="ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50" Sep 29 20:35:50 crc kubenswrapper[4741]: E0929 20:35:50.705604 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50\": container with ID starting with ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50 not found: ID does not exist" containerID="ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.705656 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50"} err="failed to get container status \"ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50\": rpc error: code = NotFound desc = could not find container \"ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50\": container with ID starting with ceeac4255846bd9f3e20e3734e4703641c9c96a6307ce41fc916ddf64a5d4b50 not found: ID does not exist" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.705687 4741 scope.go:117] "RemoveContainer" containerID="592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676" Sep 29 20:35:50 crc kubenswrapper[4741]: E0929 20:35:50.706196 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676\": container with ID starting with 592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676 not found: ID does not exist" containerID="592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676" Sep 29 20:35:50 crc kubenswrapper[4741]: I0929 20:35:50.706243 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676"} err="failed to get container status \"592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676\": rpc error: code = NotFound desc = could not find container \"592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676\": container with ID starting with 592234860bec7f6ba99f9511b2f45b8f96f2b4e37534a1437126cdd847803676 not found: ID does not exist" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.102199 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="732f829a-1021-47d7-b9b7-266405d9557f" path="/var/lib/kubelet/pods/732f829a-1021-47d7-b9b7-266405d9557f/volumes" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.583235 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.610348 4741 generic.go:334] "Generic (PLEG): container finished" podID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerID="1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557" exitCode=0 Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.610440 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btkdc" event={"ID":"20269d23-3f5c-41bb-ab7f-34eb83010a73","Type":"ContainerDied","Data":"1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557"} Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.610537 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btkdc" event={"ID":"20269d23-3f5c-41bb-ab7f-34eb83010a73","Type":"ContainerDied","Data":"62fe0c14b8336e0ffc3c4296b20c65f8f318813c9afbcc0417af8607cfd69c48"} Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.610561 4741 scope.go:117] "RemoveContainer" containerID="1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.615200 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btkdc" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.638909 4741 scope.go:117] "RemoveContainer" containerID="b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.663775 4741 scope.go:117] "RemoveContainer" containerID="3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.693582 4741 scope.go:117] "RemoveContainer" containerID="1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557" Sep 29 20:35:51 crc kubenswrapper[4741]: E0929 20:35:51.694203 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557\": container with ID starting with 1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557 not found: ID does not exist" containerID="1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.694253 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557"} err="failed to get container status \"1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557\": rpc error: code = NotFound desc = could not find container \"1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557\": container with ID starting with 1080d0b881adab607eaca98bbe7a028ae2dffcc4af44ae0079eff630956ff557 not found: ID does not exist" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.694283 4741 scope.go:117] "RemoveContainer" containerID="b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad" Sep 29 20:35:51 crc kubenswrapper[4741]: E0929 20:35:51.694685 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad\": container with ID starting with b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad not found: ID does not exist" containerID="b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.694705 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad"} err="failed to get container status \"b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad\": rpc error: code = NotFound desc = could not find container \"b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad\": container with ID starting with b86e47f77666b31ff3312eaa6ae5a7e3c4fc0b558bd997396018b44eee2e2dad not found: ID does not exist" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.694721 4741 scope.go:117] "RemoveContainer" containerID="3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc" Sep 29 20:35:51 crc kubenswrapper[4741]: E0929 20:35:51.695051 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc\": container with ID starting with 3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc not found: ID does not exist" containerID="3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.695081 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc"} err="failed to get container status \"3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc\": rpc error: code = NotFound desc = could not find container \"3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc\": container with ID starting with 3a7f5b65b6db1ad17bf6773aa805c6918c4c1b84a0ea4e0deba0183decc697cc not found: ID does not exist" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.741321 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-utilities\") pod \"20269d23-3f5c-41bb-ab7f-34eb83010a73\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.741455 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qgt2\" (UniqueName: \"kubernetes.io/projected/20269d23-3f5c-41bb-ab7f-34eb83010a73-kube-api-access-7qgt2\") pod \"20269d23-3f5c-41bb-ab7f-34eb83010a73\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.741546 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-catalog-content\") pod \"20269d23-3f5c-41bb-ab7f-34eb83010a73\" (UID: \"20269d23-3f5c-41bb-ab7f-34eb83010a73\") " Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.742631 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-utilities" (OuterVolumeSpecName: "utilities") pod "20269d23-3f5c-41bb-ab7f-34eb83010a73" (UID: "20269d23-3f5c-41bb-ab7f-34eb83010a73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.746241 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20269d23-3f5c-41bb-ab7f-34eb83010a73-kube-api-access-7qgt2" (OuterVolumeSpecName: "kube-api-access-7qgt2") pod "20269d23-3f5c-41bb-ab7f-34eb83010a73" (UID: "20269d23-3f5c-41bb-ab7f-34eb83010a73"). InnerVolumeSpecName "kube-api-access-7qgt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.815577 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20269d23-3f5c-41bb-ab7f-34eb83010a73" (UID: "20269d23-3f5c-41bb-ab7f-34eb83010a73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.843854 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.843888 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qgt2\" (UniqueName: \"kubernetes.io/projected/20269d23-3f5c-41bb-ab7f-34eb83010a73-kube-api-access-7qgt2\") on node \"crc\" DevicePath \"\"" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.843902 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20269d23-3f5c-41bb-ab7f-34eb83010a73-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.944512 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-btkdc"] Sep 29 20:35:51 crc kubenswrapper[4741]: I0929 20:35:51.952663 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-btkdc"] Sep 29 20:35:53 crc kubenswrapper[4741]: I0929 20:35:53.100044 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" path="/var/lib/kubelet/pods/20269d23-3f5c-41bb-ab7f-34eb83010a73/volumes" Sep 29 20:36:01 crc kubenswrapper[4741]: I0929 20:36:01.738797 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:36:01 crc kubenswrapper[4741]: I0929 20:36:01.739737 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:36:31 crc kubenswrapper[4741]: I0929 20:36:31.739193 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:36:31 crc kubenswrapper[4741]: I0929 20:36:31.739772 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:36:31 crc kubenswrapper[4741]: I0929 20:36:31.739813 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:36:31 crc kubenswrapper[4741]: I0929 20:36:31.740518 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:36:31 crc kubenswrapper[4741]: I0929 20:36:31.740575 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" gracePeriod=600 Sep 29 20:36:31 crc kubenswrapper[4741]: E0929 20:36:31.887958 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:36:32 crc kubenswrapper[4741]: I0929 20:36:32.018257 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" exitCode=0 Sep 29 20:36:32 crc kubenswrapper[4741]: I0929 20:36:32.018300 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4"} Sep 29 20:36:32 crc kubenswrapper[4741]: I0929 20:36:32.018338 4741 scope.go:117] "RemoveContainer" containerID="412450edb495dc305601b3802bc1fa8b0e748429e9746636a17baa95e321745c" Sep 29 20:36:32 crc kubenswrapper[4741]: I0929 20:36:32.019253 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:36:32 crc kubenswrapper[4741]: E0929 20:36:32.019537 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:36:45 crc kubenswrapper[4741]: I0929 20:36:45.086277 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:36:45 crc kubenswrapper[4741]: E0929 20:36:45.087044 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:36:58 crc kubenswrapper[4741]: I0929 20:36:58.086534 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:36:58 crc kubenswrapper[4741]: E0929 20:36:58.087353 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:37:12 crc kubenswrapper[4741]: I0929 20:37:12.087163 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:37:12 crc kubenswrapper[4741]: E0929 20:37:12.087808 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.196087 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-gdqxl"] Sep 29 20:37:18 crc kubenswrapper[4741]: E0929 20:37:18.196892 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="extract-utilities" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.196904 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="extract-utilities" Sep 29 20:37:18 crc kubenswrapper[4741]: E0929 20:37:18.196921 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="extract-content" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.196926 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="extract-content" Sep 29 20:37:18 crc kubenswrapper[4741]: E0929 20:37:18.196945 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="extract-utilities" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.196950 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="extract-utilities" Sep 29 20:37:18 crc kubenswrapper[4741]: E0929 20:37:18.196965 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="extract-content" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.196971 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="extract-content" Sep 29 20:37:18 crc kubenswrapper[4741]: E0929 20:37:18.196987 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="registry-server" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.196992 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="registry-server" Sep 29 20:37:18 crc kubenswrapper[4741]: E0929 20:37:18.197001 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="registry-server" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.197007 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="registry-server" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.197153 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="20269d23-3f5c-41bb-ab7f-34eb83010a73" containerName="registry-server" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.197172 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="732f829a-1021-47d7-b9b7-266405d9557f" containerName="registry-server" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.197685 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gdqxl" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.210154 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gdqxl"] Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.292760 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6k5h\" (UniqueName: \"kubernetes.io/projected/f440e0b7-b0d0-4db6-bc89-5173c5ec283f-kube-api-access-x6k5h\") pod \"barbican-db-create-gdqxl\" (UID: \"f440e0b7-b0d0-4db6-bc89-5173c5ec283f\") " pod="openstack/barbican-db-create-gdqxl" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.394185 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6k5h\" (UniqueName: \"kubernetes.io/projected/f440e0b7-b0d0-4db6-bc89-5173c5ec283f-kube-api-access-x6k5h\") pod \"barbican-db-create-gdqxl\" (UID: \"f440e0b7-b0d0-4db6-bc89-5173c5ec283f\") " pod="openstack/barbican-db-create-gdqxl" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.429331 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6k5h\" (UniqueName: \"kubernetes.io/projected/f440e0b7-b0d0-4db6-bc89-5173c5ec283f-kube-api-access-x6k5h\") pod \"barbican-db-create-gdqxl\" (UID: \"f440e0b7-b0d0-4db6-bc89-5173c5ec283f\") " pod="openstack/barbican-db-create-gdqxl" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.517803 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gdqxl" Sep 29 20:37:18 crc kubenswrapper[4741]: I0929 20:37:18.961458 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gdqxl"] Sep 29 20:37:19 crc kubenswrapper[4741]: I0929 20:37:19.423324 4741 generic.go:334] "Generic (PLEG): container finished" podID="f440e0b7-b0d0-4db6-bc89-5173c5ec283f" containerID="ebfdd927e0ded53243c1d4d6081b1f11b13a1601727843f2bea4af94847a9764" exitCode=0 Sep 29 20:37:19 crc kubenswrapper[4741]: I0929 20:37:19.423378 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gdqxl" event={"ID":"f440e0b7-b0d0-4db6-bc89-5173c5ec283f","Type":"ContainerDied","Data":"ebfdd927e0ded53243c1d4d6081b1f11b13a1601727843f2bea4af94847a9764"} Sep 29 20:37:19 crc kubenswrapper[4741]: I0929 20:37:19.423639 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gdqxl" event={"ID":"f440e0b7-b0d0-4db6-bc89-5173c5ec283f","Type":"ContainerStarted","Data":"a55ca409d8177e31df1fb3ba741d7ef8abf6c5cf13d432b159825bdfaddc1333"} Sep 29 20:37:20 crc kubenswrapper[4741]: I0929 20:37:20.760987 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gdqxl" Sep 29 20:37:20 crc kubenswrapper[4741]: I0929 20:37:20.832182 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6k5h\" (UniqueName: \"kubernetes.io/projected/f440e0b7-b0d0-4db6-bc89-5173c5ec283f-kube-api-access-x6k5h\") pod \"f440e0b7-b0d0-4db6-bc89-5173c5ec283f\" (UID: \"f440e0b7-b0d0-4db6-bc89-5173c5ec283f\") " Sep 29 20:37:20 crc kubenswrapper[4741]: I0929 20:37:20.842295 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f440e0b7-b0d0-4db6-bc89-5173c5ec283f-kube-api-access-x6k5h" (OuterVolumeSpecName: "kube-api-access-x6k5h") pod "f440e0b7-b0d0-4db6-bc89-5173c5ec283f" (UID: "f440e0b7-b0d0-4db6-bc89-5173c5ec283f"). InnerVolumeSpecName "kube-api-access-x6k5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:37:20 crc kubenswrapper[4741]: I0929 20:37:20.934694 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6k5h\" (UniqueName: \"kubernetes.io/projected/f440e0b7-b0d0-4db6-bc89-5173c5ec283f-kube-api-access-x6k5h\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:21 crc kubenswrapper[4741]: I0929 20:37:21.452658 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gdqxl" event={"ID":"f440e0b7-b0d0-4db6-bc89-5173c5ec283f","Type":"ContainerDied","Data":"a55ca409d8177e31df1fb3ba741d7ef8abf6c5cf13d432b159825bdfaddc1333"} Sep 29 20:37:21 crc kubenswrapper[4741]: I0929 20:37:21.452958 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a55ca409d8177e31df1fb3ba741d7ef8abf6c5cf13d432b159825bdfaddc1333" Sep 29 20:37:21 crc kubenswrapper[4741]: I0929 20:37:21.452757 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gdqxl" Sep 29 20:37:23 crc kubenswrapper[4741]: I0929 20:37:23.086441 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:37:23 crc kubenswrapper[4741]: E0929 20:37:23.087164 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.222161 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d416-account-create-m2jtt"] Sep 29 20:37:28 crc kubenswrapper[4741]: E0929 20:37:28.223148 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f440e0b7-b0d0-4db6-bc89-5173c5ec283f" containerName="mariadb-database-create" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.223174 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f440e0b7-b0d0-4db6-bc89-5173c5ec283f" containerName="mariadb-database-create" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.223466 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f440e0b7-b0d0-4db6-bc89-5173c5ec283f" containerName="mariadb-database-create" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.224291 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d416-account-create-m2jtt" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.226829 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.247511 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d416-account-create-m2jtt"] Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.378724 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6zjk\" (UniqueName: \"kubernetes.io/projected/3aaa33e1-0756-44ef-a69e-cda075bdbeb3-kube-api-access-l6zjk\") pod \"barbican-d416-account-create-m2jtt\" (UID: \"3aaa33e1-0756-44ef-a69e-cda075bdbeb3\") " pod="openstack/barbican-d416-account-create-m2jtt" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.481328 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6zjk\" (UniqueName: \"kubernetes.io/projected/3aaa33e1-0756-44ef-a69e-cda075bdbeb3-kube-api-access-l6zjk\") pod \"barbican-d416-account-create-m2jtt\" (UID: \"3aaa33e1-0756-44ef-a69e-cda075bdbeb3\") " pod="openstack/barbican-d416-account-create-m2jtt" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.514058 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6zjk\" (UniqueName: \"kubernetes.io/projected/3aaa33e1-0756-44ef-a69e-cda075bdbeb3-kube-api-access-l6zjk\") pod \"barbican-d416-account-create-m2jtt\" (UID: \"3aaa33e1-0756-44ef-a69e-cda075bdbeb3\") " pod="openstack/barbican-d416-account-create-m2jtt" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.550214 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d416-account-create-m2jtt" Sep 29 20:37:28 crc kubenswrapper[4741]: I0929 20:37:28.843353 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d416-account-create-m2jtt"] Sep 29 20:37:28 crc kubenswrapper[4741]: W0929 20:37:28.853647 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aaa33e1_0756_44ef_a69e_cda075bdbeb3.slice/crio-4459b05c2ef20df9ce079ef993fb8c919b99477de628cf8f920b00d51fb7bcbc WatchSource:0}: Error finding container 4459b05c2ef20df9ce079ef993fb8c919b99477de628cf8f920b00d51fb7bcbc: Status 404 returned error can't find the container with id 4459b05c2ef20df9ce079ef993fb8c919b99477de628cf8f920b00d51fb7bcbc Sep 29 20:37:29 crc kubenswrapper[4741]: I0929 20:37:29.526309 4741 generic.go:334] "Generic (PLEG): container finished" podID="3aaa33e1-0756-44ef-a69e-cda075bdbeb3" containerID="0db150efe72d4fbdd7c4b3c21b487efcdc61d047c0ffd28b283643d2508f52a9" exitCode=0 Sep 29 20:37:29 crc kubenswrapper[4741]: I0929 20:37:29.526375 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d416-account-create-m2jtt" event={"ID":"3aaa33e1-0756-44ef-a69e-cda075bdbeb3","Type":"ContainerDied","Data":"0db150efe72d4fbdd7c4b3c21b487efcdc61d047c0ffd28b283643d2508f52a9"} Sep 29 20:37:29 crc kubenswrapper[4741]: I0929 20:37:29.526467 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d416-account-create-m2jtt" event={"ID":"3aaa33e1-0756-44ef-a69e-cda075bdbeb3","Type":"ContainerStarted","Data":"4459b05c2ef20df9ce079ef993fb8c919b99477de628cf8f920b00d51fb7bcbc"} Sep 29 20:37:30 crc kubenswrapper[4741]: I0929 20:37:30.839709 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d416-account-create-m2jtt" Sep 29 20:37:30 crc kubenswrapper[4741]: I0929 20:37:30.924663 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6zjk\" (UniqueName: \"kubernetes.io/projected/3aaa33e1-0756-44ef-a69e-cda075bdbeb3-kube-api-access-l6zjk\") pod \"3aaa33e1-0756-44ef-a69e-cda075bdbeb3\" (UID: \"3aaa33e1-0756-44ef-a69e-cda075bdbeb3\") " Sep 29 20:37:30 crc kubenswrapper[4741]: I0929 20:37:30.929922 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aaa33e1-0756-44ef-a69e-cda075bdbeb3-kube-api-access-l6zjk" (OuterVolumeSpecName: "kube-api-access-l6zjk") pod "3aaa33e1-0756-44ef-a69e-cda075bdbeb3" (UID: "3aaa33e1-0756-44ef-a69e-cda075bdbeb3"). InnerVolumeSpecName "kube-api-access-l6zjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:37:31 crc kubenswrapper[4741]: I0929 20:37:31.026369 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6zjk\" (UniqueName: \"kubernetes.io/projected/3aaa33e1-0756-44ef-a69e-cda075bdbeb3-kube-api-access-l6zjk\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:31 crc kubenswrapper[4741]: I0929 20:37:31.547950 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d416-account-create-m2jtt" event={"ID":"3aaa33e1-0756-44ef-a69e-cda075bdbeb3","Type":"ContainerDied","Data":"4459b05c2ef20df9ce079ef993fb8c919b99477de628cf8f920b00d51fb7bcbc"} Sep 29 20:37:31 crc kubenswrapper[4741]: I0929 20:37:31.547992 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4459b05c2ef20df9ce079ef993fb8c919b99477de628cf8f920b00d51fb7bcbc" Sep 29 20:37:31 crc kubenswrapper[4741]: I0929 20:37:31.548029 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d416-account-create-m2jtt" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.443667 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-j2k9n"] Sep 29 20:37:33 crc kubenswrapper[4741]: E0929 20:37:33.444362 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aaa33e1-0756-44ef-a69e-cda075bdbeb3" containerName="mariadb-account-create" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.444380 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aaa33e1-0756-44ef-a69e-cda075bdbeb3" containerName="mariadb-account-create" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.444724 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aaa33e1-0756-44ef-a69e-cda075bdbeb3" containerName="mariadb-account-create" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.445601 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.449151 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.450193 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4lr2c" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.451660 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j2k9n"] Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.565898 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c47jq\" (UniqueName: \"kubernetes.io/projected/04834d12-fe1c-4900-b570-6097aa294692-kube-api-access-c47jq\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.566018 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-db-sync-config-data\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.566097 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-combined-ca-bundle\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.667160 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-db-sync-config-data\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.667234 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-combined-ca-bundle\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.667288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c47jq\" (UniqueName: \"kubernetes.io/projected/04834d12-fe1c-4900-b570-6097aa294692-kube-api-access-c47jq\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.671988 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-db-sync-config-data\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.690600 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c47jq\" (UniqueName: \"kubernetes.io/projected/04834d12-fe1c-4900-b570-6097aa294692-kube-api-access-c47jq\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.692144 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-combined-ca-bundle\") pod \"barbican-db-sync-j2k9n\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:33 crc kubenswrapper[4741]: I0929 20:37:33.774249 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:34 crc kubenswrapper[4741]: I0929 20:37:34.248884 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j2k9n"] Sep 29 20:37:34 crc kubenswrapper[4741]: I0929 20:37:34.575163 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j2k9n" event={"ID":"04834d12-fe1c-4900-b570-6097aa294692","Type":"ContainerStarted","Data":"c9225fd4f7e98d942096d081346767e4a082cb7c8e7dbdad44147cd1c7af0b90"} Sep 29 20:37:34 crc kubenswrapper[4741]: I0929 20:37:34.575531 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j2k9n" event={"ID":"04834d12-fe1c-4900-b570-6097aa294692","Type":"ContainerStarted","Data":"10a1c77daedf7b819dd2788688bcb299fb0ebb2deaed5c7a1e93849d122dafe9"} Sep 29 20:37:34 crc kubenswrapper[4741]: I0929 20:37:34.596641 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-j2k9n" podStartSLOduration=1.596622336 podStartE2EDuration="1.596622336s" podCreationTimestamp="2025-09-29 20:37:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:37:34.591759772 +0000 UTC m=+5296.239549104" watchObservedRunningTime="2025-09-29 20:37:34.596622336 +0000 UTC m=+5296.244411668" Sep 29 20:37:35 crc kubenswrapper[4741]: I0929 20:37:35.091044 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:37:35 crc kubenswrapper[4741]: E0929 20:37:35.091949 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:37:35 crc kubenswrapper[4741]: I0929 20:37:35.583247 4741 generic.go:334] "Generic (PLEG): container finished" podID="04834d12-fe1c-4900-b570-6097aa294692" containerID="c9225fd4f7e98d942096d081346767e4a082cb7c8e7dbdad44147cd1c7af0b90" exitCode=0 Sep 29 20:37:35 crc kubenswrapper[4741]: I0929 20:37:35.583300 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j2k9n" event={"ID":"04834d12-fe1c-4900-b570-6097aa294692","Type":"ContainerDied","Data":"c9225fd4f7e98d942096d081346767e4a082cb7c8e7dbdad44147cd1c7af0b90"} Sep 29 20:37:36 crc kubenswrapper[4741]: I0929 20:37:36.935530 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.022588 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-db-sync-config-data\") pod \"04834d12-fe1c-4900-b570-6097aa294692\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.022672 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c47jq\" (UniqueName: \"kubernetes.io/projected/04834d12-fe1c-4900-b570-6097aa294692-kube-api-access-c47jq\") pod \"04834d12-fe1c-4900-b570-6097aa294692\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.022699 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-combined-ca-bundle\") pod \"04834d12-fe1c-4900-b570-6097aa294692\" (UID: \"04834d12-fe1c-4900-b570-6097aa294692\") " Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.029449 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "04834d12-fe1c-4900-b570-6097aa294692" (UID: "04834d12-fe1c-4900-b570-6097aa294692"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.030965 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04834d12-fe1c-4900-b570-6097aa294692-kube-api-access-c47jq" (OuterVolumeSpecName: "kube-api-access-c47jq") pod "04834d12-fe1c-4900-b570-6097aa294692" (UID: "04834d12-fe1c-4900-b570-6097aa294692"). InnerVolumeSpecName "kube-api-access-c47jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.057938 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04834d12-fe1c-4900-b570-6097aa294692" (UID: "04834d12-fe1c-4900-b570-6097aa294692"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.125464 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c47jq\" (UniqueName: \"kubernetes.io/projected/04834d12-fe1c-4900-b570-6097aa294692-kube-api-access-c47jq\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.125515 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.125527 4741 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04834d12-fe1c-4900-b570-6097aa294692-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.604040 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j2k9n" event={"ID":"04834d12-fe1c-4900-b570-6097aa294692","Type":"ContainerDied","Data":"10a1c77daedf7b819dd2788688bcb299fb0ebb2deaed5c7a1e93849d122dafe9"} Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.604079 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10a1c77daedf7b819dd2788688bcb299fb0ebb2deaed5c7a1e93849d122dafe9" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.604103 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j2k9n" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.845302 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-69ddc8b4f7-ll8mc"] Sep 29 20:37:37 crc kubenswrapper[4741]: E0929 20:37:37.845655 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04834d12-fe1c-4900-b570-6097aa294692" containerName="barbican-db-sync" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.845667 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="04834d12-fe1c-4900-b570-6097aa294692" containerName="barbican-db-sync" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.845849 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="04834d12-fe1c-4900-b570-6097aa294692" containerName="barbican-db-sync" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.850535 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.856300 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.856512 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4lr2c" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.856628 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.885482 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69ddc8b4f7-ll8mc"] Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.897049 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-c97bb94fd-fj97f"] Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.899485 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.905132 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.939216 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-config-data\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.939272 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-combined-ca-bundle\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.939346 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-config-data-custom\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.939381 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804b72d0-4fb6-4e38-a52d-18703a62dbce-logs\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.939414 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv66b\" (UniqueName: \"kubernetes.io/projected/804b72d0-4fb6-4e38-a52d-18703a62dbce-kube-api-access-tv66b\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.939578 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-c97bb94fd-fj97f"] Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.965976 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-766f84f587-vnm7g"] Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.967465 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:37 crc kubenswrapper[4741]: I0929 20:37:37.978156 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766f84f587-vnm7g"] Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.040332 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-combined-ca-bundle\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.040950 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-config\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041032 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-config-data-custom\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041119 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpr7r\" (UniqueName: \"kubernetes.io/projected/bdccd1c2-21af-4189-9a3c-15fd2ea83868-kube-api-access-hpr7r\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041198 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-logs\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041282 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-config-data\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041381 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zslm8\" (UniqueName: \"kubernetes.io/projected/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-kube-api-access-zslm8\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041490 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-config-data\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041564 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-combined-ca-bundle\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041650 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-sb\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041733 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-nb\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041819 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-dns-svc\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041922 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-config-data-custom\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.041991 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804b72d0-4fb6-4e38-a52d-18703a62dbce-logs\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.042063 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv66b\" (UniqueName: \"kubernetes.io/projected/804b72d0-4fb6-4e38-a52d-18703a62dbce-kube-api-access-tv66b\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.043696 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804b72d0-4fb6-4e38-a52d-18703a62dbce-logs\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.051378 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-config-data\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.051837 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-combined-ca-bundle\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.052931 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/804b72d0-4fb6-4e38-a52d-18703a62dbce-config-data-custom\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.067081 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv66b\" (UniqueName: \"kubernetes.io/projected/804b72d0-4fb6-4e38-a52d-18703a62dbce-kube-api-access-tv66b\") pod \"barbican-worker-69ddc8b4f7-ll8mc\" (UID: \"804b72d0-4fb6-4e38-a52d-18703a62dbce\") " pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.068163 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-548cf4dc5b-fdx6w"] Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.069380 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.073256 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.091950 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-548cf4dc5b-fdx6w"] Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.143979 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zslm8\" (UniqueName: \"kubernetes.io/projected/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-kube-api-access-zslm8\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144067 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f5cab2-dd60-4453-88b6-8416abfc38e2-logs\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144103 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-sb\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144128 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-nb\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144156 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-dns-svc\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144174 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-config-data-custom\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144197 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-config-data\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144243 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-combined-ca-bundle\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144270 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-config\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144287 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rz8m\" (UniqueName: \"kubernetes.io/projected/83f5cab2-dd60-4453-88b6-8416abfc38e2-kube-api-access-4rz8m\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144307 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-config-data-custom\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144324 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpr7r\" (UniqueName: \"kubernetes.io/projected/bdccd1c2-21af-4189-9a3c-15fd2ea83868-kube-api-access-hpr7r\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144342 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-logs\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144357 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-combined-ca-bundle\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.144375 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-config-data\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.145100 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-sb\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.145901 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-config\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.146869 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-dns-svc\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.147207 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-nb\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.148922 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-logs\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.152753 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-combined-ca-bundle\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.153017 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-config-data\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.158923 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-config-data-custom\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.161438 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zslm8\" (UniqueName: \"kubernetes.io/projected/c51fb5ea-8cca-41fe-bb13-5ffeea42eb99-kube-api-access-zslm8\") pod \"barbican-keystone-listener-c97bb94fd-fj97f\" (UID: \"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99\") " pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.169020 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpr7r\" (UniqueName: \"kubernetes.io/projected/bdccd1c2-21af-4189-9a3c-15fd2ea83868-kube-api-access-hpr7r\") pod \"dnsmasq-dns-766f84f587-vnm7g\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.174831 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.229239 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.245888 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rz8m\" (UniqueName: \"kubernetes.io/projected/83f5cab2-dd60-4453-88b6-8416abfc38e2-kube-api-access-4rz8m\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.245933 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-combined-ca-bundle\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.245991 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f5cab2-dd60-4453-88b6-8416abfc38e2-logs\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.246037 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-config-data-custom\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.246057 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-config-data\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.247266 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f5cab2-dd60-4453-88b6-8416abfc38e2-logs\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.250157 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-combined-ca-bundle\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.250636 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-config-data-custom\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.263631 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f5cab2-dd60-4453-88b6-8416abfc38e2-config-data\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.280881 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rz8m\" (UniqueName: \"kubernetes.io/projected/83f5cab2-dd60-4453-88b6-8416abfc38e2-kube-api-access-4rz8m\") pod \"barbican-api-548cf4dc5b-fdx6w\" (UID: \"83f5cab2-dd60-4453-88b6-8416abfc38e2\") " pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.291863 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.430247 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.700103 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69ddc8b4f7-ll8mc"] Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.819359 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-c97bb94fd-fj97f"] Sep 29 20:37:38 crc kubenswrapper[4741]: I0929 20:37:38.926526 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766f84f587-vnm7g"] Sep 29 20:37:38 crc kubenswrapper[4741]: W0929 20:37:38.949139 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdccd1c2_21af_4189_9a3c_15fd2ea83868.slice/crio-bff9cebb2377a0d542d3dd9591b4b855759a01987048baa26f3dc6721af2b56f WatchSource:0}: Error finding container bff9cebb2377a0d542d3dd9591b4b855759a01987048baa26f3dc6721af2b56f: Status 404 returned error can't find the container with id bff9cebb2377a0d542d3dd9591b4b855759a01987048baa26f3dc6721af2b56f Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.007025 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-548cf4dc5b-fdx6w"] Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.626277 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" event={"ID":"804b72d0-4fb6-4e38-a52d-18703a62dbce","Type":"ContainerStarted","Data":"fc8808f1a5b9dfc3c19b24ae35d0fde761c9efcc00a800062194399964570739"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.626891 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" event={"ID":"804b72d0-4fb6-4e38-a52d-18703a62dbce","Type":"ContainerStarted","Data":"13864a3d452b26b7867df01ed8ce0cefed0a8d96ef2ea4fa936bdeeebc988ee5"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.626903 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" event={"ID":"804b72d0-4fb6-4e38-a52d-18703a62dbce","Type":"ContainerStarted","Data":"bb959f3b1b40073194d4c4cfe82e2e8fc810ab1a480f5e1d2754d07c03542954"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.628948 4741 generic.go:334] "Generic (PLEG): container finished" podID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerID="1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee" exitCode=0 Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.629005 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" event={"ID":"bdccd1c2-21af-4189-9a3c-15fd2ea83868","Type":"ContainerDied","Data":"1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.629096 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" event={"ID":"bdccd1c2-21af-4189-9a3c-15fd2ea83868","Type":"ContainerStarted","Data":"bff9cebb2377a0d542d3dd9591b4b855759a01987048baa26f3dc6721af2b56f"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.631617 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548cf4dc5b-fdx6w" event={"ID":"83f5cab2-dd60-4453-88b6-8416abfc38e2","Type":"ContainerStarted","Data":"9d614f2ecb995a49587a2a9461a6fb6ff33411fa3b7b49a3678d994947786095"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.631660 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548cf4dc5b-fdx6w" event={"ID":"83f5cab2-dd60-4453-88b6-8416abfc38e2","Type":"ContainerStarted","Data":"894efed5095d49356c7b56b45a3fb0a90fae92551fd9459618243f3c820cff5f"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.631670 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548cf4dc5b-fdx6w" event={"ID":"83f5cab2-dd60-4453-88b6-8416abfc38e2","Type":"ContainerStarted","Data":"03a8b61c31170280bc9407c7ddb1427bc08a2806a810b99bb4f796a77a327b95"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.631723 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.633356 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" event={"ID":"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99","Type":"ContainerStarted","Data":"632e3af39242bd5b3030bb11808ae76fb5f19ab49a687c8fce7069f9cde7d051"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.633488 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" event={"ID":"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99","Type":"ContainerStarted","Data":"e3bdc0e7f26e4abbcd6349d105350e33e3a929fa866eeb013ae8656d5d127538"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.633507 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" event={"ID":"c51fb5ea-8cca-41fe-bb13-5ffeea42eb99","Type":"ContainerStarted","Data":"317a44b896b8896fd8f074196d3b171e216795420330d65d7b5c928bc81a4214"} Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.649544 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-69ddc8b4f7-ll8mc" podStartSLOduration=2.6495237190000003 podStartE2EDuration="2.649523719s" podCreationTimestamp="2025-09-29 20:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:37:39.642433486 +0000 UTC m=+5301.290222818" watchObservedRunningTime="2025-09-29 20:37:39.649523719 +0000 UTC m=+5301.297313051" Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.668746 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-548cf4dc5b-fdx6w" podStartSLOduration=1.6687279940000002 podStartE2EDuration="1.668727994s" podCreationTimestamp="2025-09-29 20:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:37:39.666028709 +0000 UTC m=+5301.313818041" watchObservedRunningTime="2025-09-29 20:37:39.668727994 +0000 UTC m=+5301.316517326" Sep 29 20:37:39 crc kubenswrapper[4741]: I0929 20:37:39.685662 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-c97bb94fd-fj97f" podStartSLOduration=2.685646178 podStartE2EDuration="2.685646178s" podCreationTimestamp="2025-09-29 20:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:37:39.684215903 +0000 UTC m=+5301.332005235" watchObservedRunningTime="2025-09-29 20:37:39.685646178 +0000 UTC m=+5301.333435510" Sep 29 20:37:40 crc kubenswrapper[4741]: I0929 20:37:40.643759 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" event={"ID":"bdccd1c2-21af-4189-9a3c-15fd2ea83868","Type":"ContainerStarted","Data":"0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678"} Sep 29 20:37:40 crc kubenswrapper[4741]: I0929 20:37:40.644383 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:40 crc kubenswrapper[4741]: I0929 20:37:40.674666 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" podStartSLOduration=3.674648732 podStartE2EDuration="3.674648732s" podCreationTimestamp="2025-09-29 20:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:37:40.670439999 +0000 UTC m=+5302.318229341" watchObservedRunningTime="2025-09-29 20:37:40.674648732 +0000 UTC m=+5302.322438064" Sep 29 20:37:41 crc kubenswrapper[4741]: I0929 20:37:41.651779 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.087544 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:37:48 crc kubenswrapper[4741]: E0929 20:37:48.088610 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.294669 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.363532 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd4f7d747-jx7jl"] Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.364005 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" podUID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerName="dnsmasq-dns" containerID="cri-o://7cc7221bda1769b4d8634c9ddfa76c534c09e093f8ee3fbfa9abfa530f1a7325" gracePeriod=10 Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.706601 4741 generic.go:334] "Generic (PLEG): container finished" podID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerID="7cc7221bda1769b4d8634c9ddfa76c534c09e093f8ee3fbfa9abfa530f1a7325" exitCode=0 Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.706706 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" event={"ID":"174dfc5c-a80c-4ee1-84c2-427761b0b8ec","Type":"ContainerDied","Data":"7cc7221bda1769b4d8634c9ddfa76c534c09e093f8ee3fbfa9abfa530f1a7325"} Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.876799 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.932970 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-sb\") pod \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.933074 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdj8l\" (UniqueName: \"kubernetes.io/projected/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-kube-api-access-kdj8l\") pod \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.933091 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-dns-svc\") pod \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.933119 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-nb\") pod \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.933140 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-config\") pod \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\" (UID: \"174dfc5c-a80c-4ee1-84c2-427761b0b8ec\") " Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.954716 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-kube-api-access-kdj8l" (OuterVolumeSpecName: "kube-api-access-kdj8l") pod "174dfc5c-a80c-4ee1-84c2-427761b0b8ec" (UID: "174dfc5c-a80c-4ee1-84c2-427761b0b8ec"). InnerVolumeSpecName "kube-api-access-kdj8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.971152 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "174dfc5c-a80c-4ee1-84c2-427761b0b8ec" (UID: "174dfc5c-a80c-4ee1-84c2-427761b0b8ec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.971865 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "174dfc5c-a80c-4ee1-84c2-427761b0b8ec" (UID: "174dfc5c-a80c-4ee1-84c2-427761b0b8ec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.972784 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-config" (OuterVolumeSpecName: "config") pod "174dfc5c-a80c-4ee1-84c2-427761b0b8ec" (UID: "174dfc5c-a80c-4ee1-84c2-427761b0b8ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:37:48 crc kubenswrapper[4741]: I0929 20:37:48.973374 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "174dfc5c-a80c-4ee1-84c2-427761b0b8ec" (UID: "174dfc5c-a80c-4ee1-84c2-427761b0b8ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.034785 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.035038 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdj8l\" (UniqueName: \"kubernetes.io/projected/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-kube-api-access-kdj8l\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.035114 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.035197 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.035257 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/174dfc5c-a80c-4ee1-84c2-427761b0b8ec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.719880 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" event={"ID":"174dfc5c-a80c-4ee1-84c2-427761b0b8ec","Type":"ContainerDied","Data":"7ad444886e518c74fcbb73f6e329056690199d3f8640ee54c66be93f708758e0"} Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.720230 4741 scope.go:117] "RemoveContainer" containerID="7cc7221bda1769b4d8634c9ddfa76c534c09e093f8ee3fbfa9abfa530f1a7325" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.720368 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd4f7d747-jx7jl" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.757921 4741 scope.go:117] "RemoveContainer" containerID="74709b5967068c4804bc8bb6b75ca331018c117aa1f50c70e70390a1cffdbd30" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.768588 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd4f7d747-jx7jl"] Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.776850 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bd4f7d747-jx7jl"] Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.873019 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:49 crc kubenswrapper[4741]: I0929 20:37:49.945407 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-548cf4dc5b-fdx6w" Sep 29 20:37:51 crc kubenswrapper[4741]: I0929 20:37:51.095549 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" path="/var/lib/kubelet/pods/174dfc5c-a80c-4ee1-84c2-427761b0b8ec/volumes" Sep 29 20:38:02 crc kubenswrapper[4741]: I0929 20:38:02.086111 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:38:02 crc kubenswrapper[4741]: E0929 20:38:02.088038 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.451237 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-n8mwn"] Sep 29 20:38:03 crc kubenswrapper[4741]: E0929 20:38:03.451596 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerName="init" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.451609 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerName="init" Sep 29 20:38:03 crc kubenswrapper[4741]: E0929 20:38:03.451627 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerName="dnsmasq-dns" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.451633 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerName="dnsmasq-dns" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.451792 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="174dfc5c-a80c-4ee1-84c2-427761b0b8ec" containerName="dnsmasq-dns" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.452315 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-n8mwn" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.465810 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-n8mwn"] Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.488333 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9kf9\" (UniqueName: \"kubernetes.io/projected/f934fe52-4c4f-45e2-8051-a3e70052e6c6-kube-api-access-z9kf9\") pod \"neutron-db-create-n8mwn\" (UID: \"f934fe52-4c4f-45e2-8051-a3e70052e6c6\") " pod="openstack/neutron-db-create-n8mwn" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.590232 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9kf9\" (UniqueName: \"kubernetes.io/projected/f934fe52-4c4f-45e2-8051-a3e70052e6c6-kube-api-access-z9kf9\") pod \"neutron-db-create-n8mwn\" (UID: \"f934fe52-4c4f-45e2-8051-a3e70052e6c6\") " pod="openstack/neutron-db-create-n8mwn" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.607462 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9kf9\" (UniqueName: \"kubernetes.io/projected/f934fe52-4c4f-45e2-8051-a3e70052e6c6-kube-api-access-z9kf9\") pod \"neutron-db-create-n8mwn\" (UID: \"f934fe52-4c4f-45e2-8051-a3e70052e6c6\") " pod="openstack/neutron-db-create-n8mwn" Sep 29 20:38:03 crc kubenswrapper[4741]: I0929 20:38:03.776512 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-n8mwn" Sep 29 20:38:04 crc kubenswrapper[4741]: I0929 20:38:04.225200 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-n8mwn"] Sep 29 20:38:04 crc kubenswrapper[4741]: W0929 20:38:04.231079 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf934fe52_4c4f_45e2_8051_a3e70052e6c6.slice/crio-c6d80e2690d05f830fee342c6db2d988bbccee53cc8d4bca0b928de89ee4df49 WatchSource:0}: Error finding container c6d80e2690d05f830fee342c6db2d988bbccee53cc8d4bca0b928de89ee4df49: Status 404 returned error can't find the container with id c6d80e2690d05f830fee342c6db2d988bbccee53cc8d4bca0b928de89ee4df49 Sep 29 20:38:04 crc kubenswrapper[4741]: I0929 20:38:04.856238 4741 generic.go:334] "Generic (PLEG): container finished" podID="f934fe52-4c4f-45e2-8051-a3e70052e6c6" containerID="e6fda1baf2b442d1b416273464fb7067b8431bb7098fb8f6024891751dadfa51" exitCode=0 Sep 29 20:38:04 crc kubenswrapper[4741]: I0929 20:38:04.856280 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-n8mwn" event={"ID":"f934fe52-4c4f-45e2-8051-a3e70052e6c6","Type":"ContainerDied","Data":"e6fda1baf2b442d1b416273464fb7067b8431bb7098fb8f6024891751dadfa51"} Sep 29 20:38:04 crc kubenswrapper[4741]: I0929 20:38:04.856311 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-n8mwn" event={"ID":"f934fe52-4c4f-45e2-8051-a3e70052e6c6","Type":"ContainerStarted","Data":"c6d80e2690d05f830fee342c6db2d988bbccee53cc8d4bca0b928de89ee4df49"} Sep 29 20:38:06 crc kubenswrapper[4741]: I0929 20:38:06.163386 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-n8mwn" Sep 29 20:38:06 crc kubenswrapper[4741]: I0929 20:38:06.233090 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9kf9\" (UniqueName: \"kubernetes.io/projected/f934fe52-4c4f-45e2-8051-a3e70052e6c6-kube-api-access-z9kf9\") pod \"f934fe52-4c4f-45e2-8051-a3e70052e6c6\" (UID: \"f934fe52-4c4f-45e2-8051-a3e70052e6c6\") " Sep 29 20:38:06 crc kubenswrapper[4741]: I0929 20:38:06.237910 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f934fe52-4c4f-45e2-8051-a3e70052e6c6-kube-api-access-z9kf9" (OuterVolumeSpecName: "kube-api-access-z9kf9") pod "f934fe52-4c4f-45e2-8051-a3e70052e6c6" (UID: "f934fe52-4c4f-45e2-8051-a3e70052e6c6"). InnerVolumeSpecName "kube-api-access-z9kf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:38:06 crc kubenswrapper[4741]: I0929 20:38:06.335428 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9kf9\" (UniqueName: \"kubernetes.io/projected/f934fe52-4c4f-45e2-8051-a3e70052e6c6-kube-api-access-z9kf9\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:06 crc kubenswrapper[4741]: I0929 20:38:06.871674 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-n8mwn" event={"ID":"f934fe52-4c4f-45e2-8051-a3e70052e6c6","Type":"ContainerDied","Data":"c6d80e2690d05f830fee342c6db2d988bbccee53cc8d4bca0b928de89ee4df49"} Sep 29 20:38:06 crc kubenswrapper[4741]: I0929 20:38:06.871766 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6d80e2690d05f830fee342c6db2d988bbccee53cc8d4bca0b928de89ee4df49" Sep 29 20:38:06 crc kubenswrapper[4741]: I0929 20:38:06.871737 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-n8mwn" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.598827 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3c8d-account-create-4j747"] Sep 29 20:38:13 crc kubenswrapper[4741]: E0929 20:38:13.599959 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f934fe52-4c4f-45e2-8051-a3e70052e6c6" containerName="mariadb-database-create" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.599981 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f934fe52-4c4f-45e2-8051-a3e70052e6c6" containerName="mariadb-database-create" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.600220 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f934fe52-4c4f-45e2-8051-a3e70052e6c6" containerName="mariadb-database-create" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.601327 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c8d-account-create-4j747" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.607202 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3c8d-account-create-4j747"] Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.610168 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.657073 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmvts\" (UniqueName: \"kubernetes.io/projected/91cc136a-99d2-4554-80c9-1e701449f830-kube-api-access-lmvts\") pod \"neutron-3c8d-account-create-4j747\" (UID: \"91cc136a-99d2-4554-80c9-1e701449f830\") " pod="openstack/neutron-3c8d-account-create-4j747" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.759589 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmvts\" (UniqueName: \"kubernetes.io/projected/91cc136a-99d2-4554-80c9-1e701449f830-kube-api-access-lmvts\") pod \"neutron-3c8d-account-create-4j747\" (UID: \"91cc136a-99d2-4554-80c9-1e701449f830\") " pod="openstack/neutron-3c8d-account-create-4j747" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.783357 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmvts\" (UniqueName: \"kubernetes.io/projected/91cc136a-99d2-4554-80c9-1e701449f830-kube-api-access-lmvts\") pod \"neutron-3c8d-account-create-4j747\" (UID: \"91cc136a-99d2-4554-80c9-1e701449f830\") " pod="openstack/neutron-3c8d-account-create-4j747" Sep 29 20:38:13 crc kubenswrapper[4741]: I0929 20:38:13.923445 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c8d-account-create-4j747" Sep 29 20:38:14 crc kubenswrapper[4741]: I0929 20:38:14.369872 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3c8d-account-create-4j747"] Sep 29 20:38:14 crc kubenswrapper[4741]: I0929 20:38:14.941576 4741 generic.go:334] "Generic (PLEG): container finished" podID="91cc136a-99d2-4554-80c9-1e701449f830" containerID="f721e456d45383581c7a76ef22d3ad312844c3d1ccfe91422acc01d7b1ac0b5b" exitCode=0 Sep 29 20:38:14 crc kubenswrapper[4741]: I0929 20:38:14.941662 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3c8d-account-create-4j747" event={"ID":"91cc136a-99d2-4554-80c9-1e701449f830","Type":"ContainerDied","Data":"f721e456d45383581c7a76ef22d3ad312844c3d1ccfe91422acc01d7b1ac0b5b"} Sep 29 20:38:14 crc kubenswrapper[4741]: I0929 20:38:14.941858 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3c8d-account-create-4j747" event={"ID":"91cc136a-99d2-4554-80c9-1e701449f830","Type":"ContainerStarted","Data":"7dbbd95c0da5cd50431e363e7a9dae0777ae9c71d79e0db63dda9d696665fbbb"} Sep 29 20:38:15 crc kubenswrapper[4741]: I0929 20:38:15.086332 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:38:15 crc kubenswrapper[4741]: E0929 20:38:15.086578 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:38:16 crc kubenswrapper[4741]: I0929 20:38:16.279985 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c8d-account-create-4j747" Sep 29 20:38:16 crc kubenswrapper[4741]: I0929 20:38:16.345367 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmvts\" (UniqueName: \"kubernetes.io/projected/91cc136a-99d2-4554-80c9-1e701449f830-kube-api-access-lmvts\") pod \"91cc136a-99d2-4554-80c9-1e701449f830\" (UID: \"91cc136a-99d2-4554-80c9-1e701449f830\") " Sep 29 20:38:16 crc kubenswrapper[4741]: I0929 20:38:16.351076 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91cc136a-99d2-4554-80c9-1e701449f830-kube-api-access-lmvts" (OuterVolumeSpecName: "kube-api-access-lmvts") pod "91cc136a-99d2-4554-80c9-1e701449f830" (UID: "91cc136a-99d2-4554-80c9-1e701449f830"). InnerVolumeSpecName "kube-api-access-lmvts". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:38:16 crc kubenswrapper[4741]: I0929 20:38:16.448118 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmvts\" (UniqueName: \"kubernetes.io/projected/91cc136a-99d2-4554-80c9-1e701449f830-kube-api-access-lmvts\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:16 crc kubenswrapper[4741]: I0929 20:38:16.960816 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3c8d-account-create-4j747" event={"ID":"91cc136a-99d2-4554-80c9-1e701449f830","Type":"ContainerDied","Data":"7dbbd95c0da5cd50431e363e7a9dae0777ae9c71d79e0db63dda9d696665fbbb"} Sep 29 20:38:16 crc kubenswrapper[4741]: I0929 20:38:16.961186 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dbbd95c0da5cd50431e363e7a9dae0777ae9c71d79e0db63dda9d696665fbbb" Sep 29 20:38:16 crc kubenswrapper[4741]: I0929 20:38:16.960862 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3c8d-account-create-4j747" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.748777 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-wcpgv"] Sep 29 20:38:18 crc kubenswrapper[4741]: E0929 20:38:18.749189 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91cc136a-99d2-4554-80c9-1e701449f830" containerName="mariadb-account-create" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.749204 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="91cc136a-99d2-4554-80c9-1e701449f830" containerName="mariadb-account-create" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.749502 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="91cc136a-99d2-4554-80c9-1e701449f830" containerName="mariadb-account-create" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.750257 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.752986 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.753811 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fpfwx" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.754240 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.758685 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wcpgv"] Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.888911 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-combined-ca-bundle\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.888986 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-config\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.889120 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxntn\" (UniqueName: \"kubernetes.io/projected/7fc2fd00-e59e-476e-ad96-b2480837eff8-kube-api-access-mxntn\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.991330 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-combined-ca-bundle\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.991425 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-config\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:18 crc kubenswrapper[4741]: I0929 20:38:18.991544 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxntn\" (UniqueName: \"kubernetes.io/projected/7fc2fd00-e59e-476e-ad96-b2480837eff8-kube-api-access-mxntn\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:19 crc kubenswrapper[4741]: I0929 20:38:18.998163 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-config\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:19 crc kubenswrapper[4741]: I0929 20:38:18.998254 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-combined-ca-bundle\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:19 crc kubenswrapper[4741]: I0929 20:38:19.015926 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxntn\" (UniqueName: \"kubernetes.io/projected/7fc2fd00-e59e-476e-ad96-b2480837eff8-kube-api-access-mxntn\") pod \"neutron-db-sync-wcpgv\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:19 crc kubenswrapper[4741]: I0929 20:38:19.069291 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:19 crc kubenswrapper[4741]: I0929 20:38:19.553203 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wcpgv"] Sep 29 20:38:19 crc kubenswrapper[4741]: W0929 20:38:19.555234 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fc2fd00_e59e_476e_ad96_b2480837eff8.slice/crio-2366167cf6da44e10d6926087cac5210c20d0a9025c80c198fef5ddd922ac2b2 WatchSource:0}: Error finding container 2366167cf6da44e10d6926087cac5210c20d0a9025c80c198fef5ddd922ac2b2: Status 404 returned error can't find the container with id 2366167cf6da44e10d6926087cac5210c20d0a9025c80c198fef5ddd922ac2b2 Sep 29 20:38:19 crc kubenswrapper[4741]: I0929 20:38:19.982260 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wcpgv" event={"ID":"7fc2fd00-e59e-476e-ad96-b2480837eff8","Type":"ContainerStarted","Data":"e29d5d1ad083a20d0549b54e615438c618b3fc3ffdfe38f0bed341ea57ad555b"} Sep 29 20:38:19 crc kubenswrapper[4741]: I0929 20:38:19.982569 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wcpgv" event={"ID":"7fc2fd00-e59e-476e-ad96-b2480837eff8","Type":"ContainerStarted","Data":"2366167cf6da44e10d6926087cac5210c20d0a9025c80c198fef5ddd922ac2b2"} Sep 29 20:38:20 crc kubenswrapper[4741]: I0929 20:38:20.001510 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-wcpgv" podStartSLOduration=2.00148765 podStartE2EDuration="2.00148765s" podCreationTimestamp="2025-09-29 20:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:38:19.999515019 +0000 UTC m=+5341.647304381" watchObservedRunningTime="2025-09-29 20:38:20.00148765 +0000 UTC m=+5341.649276982" Sep 29 20:38:24 crc kubenswrapper[4741]: I0929 20:38:24.016053 4741 generic.go:334] "Generic (PLEG): container finished" podID="7fc2fd00-e59e-476e-ad96-b2480837eff8" containerID="e29d5d1ad083a20d0549b54e615438c618b3fc3ffdfe38f0bed341ea57ad555b" exitCode=0 Sep 29 20:38:24 crc kubenswrapper[4741]: I0929 20:38:24.016097 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wcpgv" event={"ID":"7fc2fd00-e59e-476e-ad96-b2480837eff8","Type":"ContainerDied","Data":"e29d5d1ad083a20d0549b54e615438c618b3fc3ffdfe38f0bed341ea57ad555b"} Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.320200 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.420597 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-combined-ca-bundle\") pod \"7fc2fd00-e59e-476e-ad96-b2480837eff8\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.420969 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxntn\" (UniqueName: \"kubernetes.io/projected/7fc2fd00-e59e-476e-ad96-b2480837eff8-kube-api-access-mxntn\") pod \"7fc2fd00-e59e-476e-ad96-b2480837eff8\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.421132 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-config\") pod \"7fc2fd00-e59e-476e-ad96-b2480837eff8\" (UID: \"7fc2fd00-e59e-476e-ad96-b2480837eff8\") " Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.426240 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc2fd00-e59e-476e-ad96-b2480837eff8-kube-api-access-mxntn" (OuterVolumeSpecName: "kube-api-access-mxntn") pod "7fc2fd00-e59e-476e-ad96-b2480837eff8" (UID: "7fc2fd00-e59e-476e-ad96-b2480837eff8"). InnerVolumeSpecName "kube-api-access-mxntn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.443179 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-config" (OuterVolumeSpecName: "config") pod "7fc2fd00-e59e-476e-ad96-b2480837eff8" (UID: "7fc2fd00-e59e-476e-ad96-b2480837eff8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.443575 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fc2fd00-e59e-476e-ad96-b2480837eff8" (UID: "7fc2fd00-e59e-476e-ad96-b2480837eff8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.523478 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.523524 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc2fd00-e59e-476e-ad96-b2480837eff8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:25 crc kubenswrapper[4741]: I0929 20:38:25.523536 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxntn\" (UniqueName: \"kubernetes.io/projected/7fc2fd00-e59e-476e-ad96-b2480837eff8-kube-api-access-mxntn\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.034719 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wcpgv" event={"ID":"7fc2fd00-e59e-476e-ad96-b2480837eff8","Type":"ContainerDied","Data":"2366167cf6da44e10d6926087cac5210c20d0a9025c80c198fef5ddd922ac2b2"} Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.035014 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2366167cf6da44e10d6926087cac5210c20d0a9025c80c198fef5ddd922ac2b2" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.034806 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wcpgv" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.180084 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8677bf69bf-c28js"] Sep 29 20:38:26 crc kubenswrapper[4741]: E0929 20:38:26.180575 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc2fd00-e59e-476e-ad96-b2480837eff8" containerName="neutron-db-sync" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.180596 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc2fd00-e59e-476e-ad96-b2480837eff8" containerName="neutron-db-sync" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.180829 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc2fd00-e59e-476e-ad96-b2480837eff8" containerName="neutron-db-sync" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.182220 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.195462 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8677bf69bf-c28js"] Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.312951 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ccc98d775-csfpc"] Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.314266 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.317258 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.318112 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.318676 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fpfwx" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.332730 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ccc98d775-csfpc"] Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.334084 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl8x2\" (UniqueName: \"kubernetes.io/projected/0a3a22ac-699b-492e-85ad-3fb34c5697fa-kube-api-access-tl8x2\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.334126 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-dns-svc\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.334194 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-sb\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.334222 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-nb\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.334410 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-config\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.435714 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-sb\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.435783 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-combined-ca-bundle\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.435809 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-nb\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.435840 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-httpd-config\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.435857 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ngv7\" (UniqueName: \"kubernetes.io/projected/1a4e588c-5749-4258-9986-b32c09909091-kube-api-access-9ngv7\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.435968 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-config\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.436076 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl8x2\" (UniqueName: \"kubernetes.io/projected/0a3a22ac-699b-492e-85ad-3fb34c5697fa-kube-api-access-tl8x2\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.436143 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-dns-svc\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.436268 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-config\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.436715 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-nb\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.436741 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-sb\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.437076 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-config\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.437349 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-dns-svc\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.454818 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl8x2\" (UniqueName: \"kubernetes.io/projected/0a3a22ac-699b-492e-85ad-3fb34c5697fa-kube-api-access-tl8x2\") pod \"dnsmasq-dns-8677bf69bf-c28js\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.508452 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.540302 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-config\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.540374 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-combined-ca-bundle\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.540417 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-httpd-config\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.540434 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ngv7\" (UniqueName: \"kubernetes.io/projected/1a4e588c-5749-4258-9986-b32c09909091-kube-api-access-9ngv7\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.544862 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-combined-ca-bundle\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.545051 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-config\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.545131 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a4e588c-5749-4258-9986-b32c09909091-httpd-config\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.558027 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ngv7\" (UniqueName: \"kubernetes.io/projected/1a4e588c-5749-4258-9986-b32c09909091-kube-api-access-9ngv7\") pod \"neutron-ccc98d775-csfpc\" (UID: \"1a4e588c-5749-4258-9986-b32c09909091\") " pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.637196 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:26 crc kubenswrapper[4741]: I0929 20:38:26.788219 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8677bf69bf-c28js"] Sep 29 20:38:27 crc kubenswrapper[4741]: I0929 20:38:27.042888 4741 generic.go:334] "Generic (PLEG): container finished" podID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerID="c8998e3bdbea1361f89ecb91adf4d07d8f21ea00dbc3b3c5127a283f0632779d" exitCode=0 Sep 29 20:38:27 crc kubenswrapper[4741]: I0929 20:38:27.043239 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" event={"ID":"0a3a22ac-699b-492e-85ad-3fb34c5697fa","Type":"ContainerDied","Data":"c8998e3bdbea1361f89ecb91adf4d07d8f21ea00dbc3b3c5127a283f0632779d"} Sep 29 20:38:27 crc kubenswrapper[4741]: I0929 20:38:27.043330 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" event={"ID":"0a3a22ac-699b-492e-85ad-3fb34c5697fa","Type":"ContainerStarted","Data":"bf61d0c79230326598e99e167891b4d7bbc40833995f4e24cc8e76e915e0aeff"} Sep 29 20:38:27 crc kubenswrapper[4741]: I0929 20:38:27.195117 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ccc98d775-csfpc"] Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.052425 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccc98d775-csfpc" event={"ID":"1a4e588c-5749-4258-9986-b32c09909091","Type":"ContainerStarted","Data":"6f7f2ec4fed3aecb25b2e83928aff117bc4b3d4947289cdc08c2690e1078abac"} Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.052798 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.052813 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccc98d775-csfpc" event={"ID":"1a4e588c-5749-4258-9986-b32c09909091","Type":"ContainerStarted","Data":"a561e0a233aa091a50286ce1bf617e54b19e4c3eedaae498c401d61a9393f19b"} Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.052826 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccc98d775-csfpc" event={"ID":"1a4e588c-5749-4258-9986-b32c09909091","Type":"ContainerStarted","Data":"ecc9f796aa30278ba5abb30a6ad7c92ca28abb1a1c3c02d106b7e8229b12539a"} Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.055156 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" event={"ID":"0a3a22ac-699b-492e-85ad-3fb34c5697fa","Type":"ContainerStarted","Data":"835cc77c0151c98a90f4e774cd7c9e02b4afdba9bfa9bdb89fbf49e24d301acd"} Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.055358 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.066974 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ccc98d775-csfpc" podStartSLOduration=2.066954619 podStartE2EDuration="2.066954619s" podCreationTimestamp="2025-09-29 20:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:38:28.06536447 +0000 UTC m=+5349.713153812" watchObservedRunningTime="2025-09-29 20:38:28.066954619 +0000 UTC m=+5349.714743951" Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.086355 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:38:28 crc kubenswrapper[4741]: E0929 20:38:28.086662 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:38:28 crc kubenswrapper[4741]: I0929 20:38:28.088345 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" podStartSLOduration=2.088333322 podStartE2EDuration="2.088333322s" podCreationTimestamp="2025-09-29 20:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:38:28.0860132 +0000 UTC m=+5349.733802542" watchObservedRunningTime="2025-09-29 20:38:28.088333322 +0000 UTC m=+5349.736122654" Sep 29 20:38:36 crc kubenswrapper[4741]: I0929 20:38:36.510600 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:38:36 crc kubenswrapper[4741]: I0929 20:38:36.618999 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766f84f587-vnm7g"] Sep 29 20:38:36 crc kubenswrapper[4741]: I0929 20:38:36.619230 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" podUID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerName="dnsmasq-dns" containerID="cri-o://0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678" gracePeriod=10 Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.074155 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.134162 4741 generic.go:334] "Generic (PLEG): container finished" podID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerID="0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678" exitCode=0 Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.134209 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" event={"ID":"bdccd1c2-21af-4189-9a3c-15fd2ea83868","Type":"ContainerDied","Data":"0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678"} Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.134253 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" event={"ID":"bdccd1c2-21af-4189-9a3c-15fd2ea83868","Type":"ContainerDied","Data":"bff9cebb2377a0d542d3dd9591b4b855759a01987048baa26f3dc6721af2b56f"} Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.134273 4741 scope.go:117] "RemoveContainer" containerID="0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.134470 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f84f587-vnm7g" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.152023 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-config\") pod \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.152074 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-dns-svc\") pod \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.152222 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-nb\") pod \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.152410 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpr7r\" (UniqueName: \"kubernetes.io/projected/bdccd1c2-21af-4189-9a3c-15fd2ea83868-kube-api-access-hpr7r\") pod \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.152519 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-sb\") pod \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\" (UID: \"bdccd1c2-21af-4189-9a3c-15fd2ea83868\") " Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.164700 4741 scope.go:117] "RemoveContainer" containerID="1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.164768 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdccd1c2-21af-4189-9a3c-15fd2ea83868-kube-api-access-hpr7r" (OuterVolumeSpecName: "kube-api-access-hpr7r") pod "bdccd1c2-21af-4189-9a3c-15fd2ea83868" (UID: "bdccd1c2-21af-4189-9a3c-15fd2ea83868"). InnerVolumeSpecName "kube-api-access-hpr7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.192904 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bdccd1c2-21af-4189-9a3c-15fd2ea83868" (UID: "bdccd1c2-21af-4189-9a3c-15fd2ea83868"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.195513 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bdccd1c2-21af-4189-9a3c-15fd2ea83868" (UID: "bdccd1c2-21af-4189-9a3c-15fd2ea83868"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.197086 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bdccd1c2-21af-4189-9a3c-15fd2ea83868" (UID: "bdccd1c2-21af-4189-9a3c-15fd2ea83868"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.214697 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-config" (OuterVolumeSpecName: "config") pod "bdccd1c2-21af-4189-9a3c-15fd2ea83868" (UID: "bdccd1c2-21af-4189-9a3c-15fd2ea83868"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.250063 4741 scope.go:117] "RemoveContainer" containerID="0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678" Sep 29 20:38:37 crc kubenswrapper[4741]: E0929 20:38:37.250621 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678\": container with ID starting with 0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678 not found: ID does not exist" containerID="0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.250675 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678"} err="failed to get container status \"0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678\": rpc error: code = NotFound desc = could not find container \"0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678\": container with ID starting with 0521e0cd1363e3ead79cc19ca3289f35774107dd1b6c71cbfb7cd53cf846f678 not found: ID does not exist" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.250716 4741 scope.go:117] "RemoveContainer" containerID="1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee" Sep 29 20:38:37 crc kubenswrapper[4741]: E0929 20:38:37.251136 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee\": container with ID starting with 1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee not found: ID does not exist" containerID="1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.251163 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee"} err="failed to get container status \"1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee\": rpc error: code = NotFound desc = could not find container \"1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee\": container with ID starting with 1fe4f300c271aa70c16f0a9e2a74e52707772d84e18aaacffdffeceee06850ee not found: ID does not exist" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.254785 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpr7r\" (UniqueName: \"kubernetes.io/projected/bdccd1c2-21af-4189-9a3c-15fd2ea83868-kube-api-access-hpr7r\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.254813 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.254868 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.254879 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.254890 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdccd1c2-21af-4189-9a3c-15fd2ea83868-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.473968 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766f84f587-vnm7g"] Sep 29 20:38:37 crc kubenswrapper[4741]: I0929 20:38:37.482551 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-766f84f587-vnm7g"] Sep 29 20:38:39 crc kubenswrapper[4741]: I0929 20:38:39.102354 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" path="/var/lib/kubelet/pods/bdccd1c2-21af-4189-9a3c-15fd2ea83868/volumes" Sep 29 20:38:43 crc kubenswrapper[4741]: I0929 20:38:43.086882 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:38:43 crc kubenswrapper[4741]: E0929 20:38:43.087752 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:38:55 crc kubenswrapper[4741]: I0929 20:38:55.086155 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:38:55 crc kubenswrapper[4741]: E0929 20:38:55.086756 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:38:56 crc kubenswrapper[4741]: I0929 20:38:56.657456 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-ccc98d775-csfpc" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.580443 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zln86"] Sep 29 20:39:04 crc kubenswrapper[4741]: E0929 20:39:04.581423 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerName="dnsmasq-dns" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.581441 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerName="dnsmasq-dns" Sep 29 20:39:04 crc kubenswrapper[4741]: E0929 20:39:04.581459 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerName="init" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.581467 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerName="init" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.581678 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdccd1c2-21af-4189-9a3c-15fd2ea83868" containerName="dnsmasq-dns" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.582493 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zln86" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.589312 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zln86"] Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.746495 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bm9\" (UniqueName: \"kubernetes.io/projected/dfbf2271-9d58-45c6-9e0b-961036c2c627-kube-api-access-99bm9\") pod \"glance-db-create-zln86\" (UID: \"dfbf2271-9d58-45c6-9e0b-961036c2c627\") " pod="openstack/glance-db-create-zln86" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.847766 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bm9\" (UniqueName: \"kubernetes.io/projected/dfbf2271-9d58-45c6-9e0b-961036c2c627-kube-api-access-99bm9\") pod \"glance-db-create-zln86\" (UID: \"dfbf2271-9d58-45c6-9e0b-961036c2c627\") " pod="openstack/glance-db-create-zln86" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.866374 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bm9\" (UniqueName: \"kubernetes.io/projected/dfbf2271-9d58-45c6-9e0b-961036c2c627-kube-api-access-99bm9\") pod \"glance-db-create-zln86\" (UID: \"dfbf2271-9d58-45c6-9e0b-961036c2c627\") " pod="openstack/glance-db-create-zln86" Sep 29 20:39:04 crc kubenswrapper[4741]: I0929 20:39:04.935020 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zln86" Sep 29 20:39:05 crc kubenswrapper[4741]: I0929 20:39:05.360650 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zln86"] Sep 29 20:39:05 crc kubenswrapper[4741]: W0929 20:39:05.370818 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfbf2271_9d58_45c6_9e0b_961036c2c627.slice/crio-1c578e7da6329c1e449290f6d8ba517fe99d09ce388e34a843bdbefbe86e9791 WatchSource:0}: Error finding container 1c578e7da6329c1e449290f6d8ba517fe99d09ce388e34a843bdbefbe86e9791: Status 404 returned error can't find the container with id 1c578e7da6329c1e449290f6d8ba517fe99d09ce388e34a843bdbefbe86e9791 Sep 29 20:39:06 crc kubenswrapper[4741]: I0929 20:39:06.085281 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:39:06 crc kubenswrapper[4741]: E0929 20:39:06.085854 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:39:06 crc kubenswrapper[4741]: I0929 20:39:06.396217 4741 generic.go:334] "Generic (PLEG): container finished" podID="dfbf2271-9d58-45c6-9e0b-961036c2c627" containerID="ced716cb51e4174f4da4df1c9b3e22239cb5f0db87b81e7c5bb10148cfaa27b8" exitCode=0 Sep 29 20:39:06 crc kubenswrapper[4741]: I0929 20:39:06.396269 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zln86" event={"ID":"dfbf2271-9d58-45c6-9e0b-961036c2c627","Type":"ContainerDied","Data":"ced716cb51e4174f4da4df1c9b3e22239cb5f0db87b81e7c5bb10148cfaa27b8"} Sep 29 20:39:06 crc kubenswrapper[4741]: I0929 20:39:06.396306 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zln86" event={"ID":"dfbf2271-9d58-45c6-9e0b-961036c2c627","Type":"ContainerStarted","Data":"1c578e7da6329c1e449290f6d8ba517fe99d09ce388e34a843bdbefbe86e9791"} Sep 29 20:39:07 crc kubenswrapper[4741]: I0929 20:39:07.772704 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zln86" Sep 29 20:39:07 crc kubenswrapper[4741]: I0929 20:39:07.895245 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bm9\" (UniqueName: \"kubernetes.io/projected/dfbf2271-9d58-45c6-9e0b-961036c2c627-kube-api-access-99bm9\") pod \"dfbf2271-9d58-45c6-9e0b-961036c2c627\" (UID: \"dfbf2271-9d58-45c6-9e0b-961036c2c627\") " Sep 29 20:39:07 crc kubenswrapper[4741]: I0929 20:39:07.902227 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfbf2271-9d58-45c6-9e0b-961036c2c627-kube-api-access-99bm9" (OuterVolumeSpecName: "kube-api-access-99bm9") pod "dfbf2271-9d58-45c6-9e0b-961036c2c627" (UID: "dfbf2271-9d58-45c6-9e0b-961036c2c627"). InnerVolumeSpecName "kube-api-access-99bm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:07 crc kubenswrapper[4741]: I0929 20:39:07.998183 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bm9\" (UniqueName: \"kubernetes.io/projected/dfbf2271-9d58-45c6-9e0b-961036c2c627-kube-api-access-99bm9\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:08 crc kubenswrapper[4741]: I0929 20:39:08.417725 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zln86" event={"ID":"dfbf2271-9d58-45c6-9e0b-961036c2c627","Type":"ContainerDied","Data":"1c578e7da6329c1e449290f6d8ba517fe99d09ce388e34a843bdbefbe86e9791"} Sep 29 20:39:08 crc kubenswrapper[4741]: I0929 20:39:08.417771 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c578e7da6329c1e449290f6d8ba517fe99d09ce388e34a843bdbefbe86e9791" Sep 29 20:39:08 crc kubenswrapper[4741]: I0929 20:39:08.417809 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zln86" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.654344 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-4c9d-account-create-f9lfc"] Sep 29 20:39:14 crc kubenswrapper[4741]: E0929 20:39:14.656468 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbf2271-9d58-45c6-9e0b-961036c2c627" containerName="mariadb-database-create" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.656606 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbf2271-9d58-45c6-9e0b-961036c2c627" containerName="mariadb-database-create" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.657104 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfbf2271-9d58-45c6-9e0b-961036c2c627" containerName="mariadb-database-create" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.658239 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c9d-account-create-f9lfc" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.660366 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.668080 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4c9d-account-create-f9lfc"] Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.820686 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gl6r\" (UniqueName: \"kubernetes.io/projected/37b6532e-5bff-48f4-8b82-78d6eb89eeba-kube-api-access-8gl6r\") pod \"glance-4c9d-account-create-f9lfc\" (UID: \"37b6532e-5bff-48f4-8b82-78d6eb89eeba\") " pod="openstack/glance-4c9d-account-create-f9lfc" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.922016 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gl6r\" (UniqueName: \"kubernetes.io/projected/37b6532e-5bff-48f4-8b82-78d6eb89eeba-kube-api-access-8gl6r\") pod \"glance-4c9d-account-create-f9lfc\" (UID: \"37b6532e-5bff-48f4-8b82-78d6eb89eeba\") " pod="openstack/glance-4c9d-account-create-f9lfc" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.941514 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gl6r\" (UniqueName: \"kubernetes.io/projected/37b6532e-5bff-48f4-8b82-78d6eb89eeba-kube-api-access-8gl6r\") pod \"glance-4c9d-account-create-f9lfc\" (UID: \"37b6532e-5bff-48f4-8b82-78d6eb89eeba\") " pod="openstack/glance-4c9d-account-create-f9lfc" Sep 29 20:39:14 crc kubenswrapper[4741]: I0929 20:39:14.990653 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c9d-account-create-f9lfc" Sep 29 20:39:15 crc kubenswrapper[4741]: I0929 20:39:15.419340 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4c9d-account-create-f9lfc"] Sep 29 20:39:15 crc kubenswrapper[4741]: I0929 20:39:15.479752 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4c9d-account-create-f9lfc" event={"ID":"37b6532e-5bff-48f4-8b82-78d6eb89eeba","Type":"ContainerStarted","Data":"cb72cc05421f42e15d66215ff16c28f859d9a8d8b1f4d3fae81b5861a8d23e1a"} Sep 29 20:39:16 crc kubenswrapper[4741]: I0929 20:39:16.490254 4741 generic.go:334] "Generic (PLEG): container finished" podID="37b6532e-5bff-48f4-8b82-78d6eb89eeba" containerID="c763d816c0f59d938e26f8ec0ba6aff3a57c1a8a012e795e7797d144cbc1f8ea" exitCode=0 Sep 29 20:39:16 crc kubenswrapper[4741]: I0929 20:39:16.490645 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4c9d-account-create-f9lfc" event={"ID":"37b6532e-5bff-48f4-8b82-78d6eb89eeba","Type":"ContainerDied","Data":"c763d816c0f59d938e26f8ec0ba6aff3a57c1a8a012e795e7797d144cbc1f8ea"} Sep 29 20:39:17 crc kubenswrapper[4741]: I0929 20:39:17.832341 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c9d-account-create-f9lfc" Sep 29 20:39:17 crc kubenswrapper[4741]: I0929 20:39:17.977006 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gl6r\" (UniqueName: \"kubernetes.io/projected/37b6532e-5bff-48f4-8b82-78d6eb89eeba-kube-api-access-8gl6r\") pod \"37b6532e-5bff-48f4-8b82-78d6eb89eeba\" (UID: \"37b6532e-5bff-48f4-8b82-78d6eb89eeba\") " Sep 29 20:39:17 crc kubenswrapper[4741]: I0929 20:39:17.983977 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b6532e-5bff-48f4-8b82-78d6eb89eeba-kube-api-access-8gl6r" (OuterVolumeSpecName: "kube-api-access-8gl6r") pod "37b6532e-5bff-48f4-8b82-78d6eb89eeba" (UID: "37b6532e-5bff-48f4-8b82-78d6eb89eeba"). InnerVolumeSpecName "kube-api-access-8gl6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:18 crc kubenswrapper[4741]: I0929 20:39:18.078894 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gl6r\" (UniqueName: \"kubernetes.io/projected/37b6532e-5bff-48f4-8b82-78d6eb89eeba-kube-api-access-8gl6r\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:18 crc kubenswrapper[4741]: I0929 20:39:18.505779 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4c9d-account-create-f9lfc" event={"ID":"37b6532e-5bff-48f4-8b82-78d6eb89eeba","Type":"ContainerDied","Data":"cb72cc05421f42e15d66215ff16c28f859d9a8d8b1f4d3fae81b5861a8d23e1a"} Sep 29 20:39:18 crc kubenswrapper[4741]: I0929 20:39:18.505825 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb72cc05421f42e15d66215ff16c28f859d9a8d8b1f4d3fae81b5861a8d23e1a" Sep 29 20:39:18 crc kubenswrapper[4741]: I0929 20:39:18.505830 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c9d-account-create-f9lfc" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.742284 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-rbmn8"] Sep 29 20:39:19 crc kubenswrapper[4741]: E0929 20:39:19.743032 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b6532e-5bff-48f4-8b82-78d6eb89eeba" containerName="mariadb-account-create" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.743062 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b6532e-5bff-48f4-8b82-78d6eb89eeba" containerName="mariadb-account-create" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.743292 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b6532e-5bff-48f4-8b82-78d6eb89eeba" containerName="mariadb-account-create" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.743977 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.751691 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.751884 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4b8tf" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.753199 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rbmn8"] Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.913268 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-487k9\" (UniqueName: \"kubernetes.io/projected/f6558a7e-b784-4232-999e-657db5dccd23-kube-api-access-487k9\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.913553 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-combined-ca-bundle\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.913762 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-db-sync-config-data\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:19 crc kubenswrapper[4741]: I0929 20:39:19.913825 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-config-data\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.015055 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-db-sync-config-data\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.015108 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-config-data\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.015186 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-487k9\" (UniqueName: \"kubernetes.io/projected/f6558a7e-b784-4232-999e-657db5dccd23-kube-api-access-487k9\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.015253 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-combined-ca-bundle\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.020370 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-combined-ca-bundle\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.021614 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-db-sync-config-data\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.031481 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-config-data\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.031767 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-487k9\" (UniqueName: \"kubernetes.io/projected/f6558a7e-b784-4232-999e-657db5dccd23-kube-api-access-487k9\") pod \"glance-db-sync-rbmn8\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.075297 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.085759 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:39:20 crc kubenswrapper[4741]: E0929 20:39:20.085985 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.439730 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rbmn8"] Sep 29 20:39:20 crc kubenswrapper[4741]: I0929 20:39:20.524859 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbmn8" event={"ID":"f6558a7e-b784-4232-999e-657db5dccd23","Type":"ContainerStarted","Data":"8770e2fc9dc04d1e3750afa6676cdcc0f6d129f2a2e9c5e902dc7bfd97b4d279"} Sep 29 20:39:21 crc kubenswrapper[4741]: I0929 20:39:21.532952 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbmn8" event={"ID":"f6558a7e-b784-4232-999e-657db5dccd23","Type":"ContainerStarted","Data":"262f21044f2140d3cf9060091c75dec8287e72aeeca82e207e092274a3428c8e"} Sep 29 20:39:21 crc kubenswrapper[4741]: I0929 20:39:21.557281 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-rbmn8" podStartSLOduration=2.557259234 podStartE2EDuration="2.557259234s" podCreationTimestamp="2025-09-29 20:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:39:21.552547048 +0000 UTC m=+5403.200336400" watchObservedRunningTime="2025-09-29 20:39:21.557259234 +0000 UTC m=+5403.205048566" Sep 29 20:39:24 crc kubenswrapper[4741]: I0929 20:39:24.559683 4741 generic.go:334] "Generic (PLEG): container finished" podID="f6558a7e-b784-4232-999e-657db5dccd23" containerID="262f21044f2140d3cf9060091c75dec8287e72aeeca82e207e092274a3428c8e" exitCode=0 Sep 29 20:39:24 crc kubenswrapper[4741]: I0929 20:39:24.559780 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbmn8" event={"ID":"f6558a7e-b784-4232-999e-657db5dccd23","Type":"ContainerDied","Data":"262f21044f2140d3cf9060091c75dec8287e72aeeca82e207e092274a3428c8e"} Sep 29 20:39:25 crc kubenswrapper[4741]: I0929 20:39:25.976163 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.031301 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-487k9\" (UniqueName: \"kubernetes.io/projected/f6558a7e-b784-4232-999e-657db5dccd23-kube-api-access-487k9\") pod \"f6558a7e-b784-4232-999e-657db5dccd23\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.031372 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-combined-ca-bundle\") pod \"f6558a7e-b784-4232-999e-657db5dccd23\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.031487 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-db-sync-config-data\") pod \"f6558a7e-b784-4232-999e-657db5dccd23\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.031516 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-config-data\") pod \"f6558a7e-b784-4232-999e-657db5dccd23\" (UID: \"f6558a7e-b784-4232-999e-657db5dccd23\") " Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.037627 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f6558a7e-b784-4232-999e-657db5dccd23" (UID: "f6558a7e-b784-4232-999e-657db5dccd23"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.037672 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6558a7e-b784-4232-999e-657db5dccd23-kube-api-access-487k9" (OuterVolumeSpecName: "kube-api-access-487k9") pod "f6558a7e-b784-4232-999e-657db5dccd23" (UID: "f6558a7e-b784-4232-999e-657db5dccd23"). InnerVolumeSpecName "kube-api-access-487k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.057639 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6558a7e-b784-4232-999e-657db5dccd23" (UID: "f6558a7e-b784-4232-999e-657db5dccd23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.079447 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-config-data" (OuterVolumeSpecName: "config-data") pod "f6558a7e-b784-4232-999e-657db5dccd23" (UID: "f6558a7e-b784-4232-999e-657db5dccd23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.134182 4741 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.134226 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.134241 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-487k9\" (UniqueName: \"kubernetes.io/projected/f6558a7e-b784-4232-999e-657db5dccd23-kube-api-access-487k9\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.134252 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6558a7e-b784-4232-999e-657db5dccd23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.575600 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbmn8" event={"ID":"f6558a7e-b784-4232-999e-657db5dccd23","Type":"ContainerDied","Data":"8770e2fc9dc04d1e3750afa6676cdcc0f6d129f2a2e9c5e902dc7bfd97b4d279"} Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.575639 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8770e2fc9dc04d1e3750afa6676cdcc0f6d129f2a2e9c5e902dc7bfd97b4d279" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.575669 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbmn8" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.955755 4741 scope.go:117] "RemoveContainer" containerID="048337c606cc646b0647f93462eb6f7d95be2a851f05d554430c7c9f4c5b9051" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.986553 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:26 crc kubenswrapper[4741]: E0929 20:39:26.987033 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6558a7e-b784-4232-999e-657db5dccd23" containerName="glance-db-sync" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.987051 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6558a7e-b784-4232-999e-657db5dccd23" containerName="glance-db-sync" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.987250 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6558a7e-b784-4232-999e-657db5dccd23" containerName="glance-db-sync" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.988429 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.991008 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.991360 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.994954 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 20:39:26 crc kubenswrapper[4741]: I0929 20:39:26.995131 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4b8tf" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.015861 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.032289 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-744dc97445-v5xqc"] Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.033717 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.061214 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.061512 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.061689 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hjpk\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-kube-api-access-5hjpk\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.061800 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-config-data\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.061997 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-logs\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.062138 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-scripts\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.062261 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-ceph\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.062852 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744dc97445-v5xqc"] Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.164927 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-config\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.164998 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-logs\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165038 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-sb\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165095 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-scripts\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165248 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-ceph\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165341 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165677 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-nb\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165725 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7dck\" (UniqueName: \"kubernetes.io/projected/217dfc37-f84b-4114-b2be-b2b55b67ce92-kube-api-access-h7dck\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165766 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-logs\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165788 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165913 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-dns-svc\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.165959 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hjpk\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-kube-api-access-5hjpk\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.166022 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-config-data\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.168765 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.170919 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-ceph\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.174256 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.183983 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-scripts\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.184723 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-config-data\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.188055 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hjpk\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-kube-api-access-5hjpk\") pod \"glance-default-external-api-0\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.190453 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.192252 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.196602 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.199940 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267791 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-logs\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267846 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-config\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267872 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267894 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-sb\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267939 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267957 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267978 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-nb\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.267997 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7dck\" (UniqueName: \"kubernetes.io/projected/217dfc37-f84b-4114-b2be-b2b55b67ce92-kube-api-access-h7dck\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.268013 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmqbb\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-kube-api-access-wmqbb\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.268044 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-dns-svc\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.268070 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.268106 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.269230 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-config\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.269346 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-nb\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.269878 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-dns-svc\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.270066 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-sb\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.298229 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7dck\" (UniqueName: \"kubernetes.io/projected/217dfc37-f84b-4114-b2be-b2b55b67ce92-kube-api-access-h7dck\") pod \"dnsmasq-dns-744dc97445-v5xqc\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.331851 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.367794 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.369413 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.369461 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-logs\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.369537 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.369631 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.369685 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.369743 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmqbb\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-kube-api-access-wmqbb\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.369815 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.372820 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.373103 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-logs\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.373663 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.374313 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.380692 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.380863 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.391143 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmqbb\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-kube-api-access-wmqbb\") pod \"glance-default-internal-api-0\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.576942 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:27 crc kubenswrapper[4741]: I0929 20:39:27.747726 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744dc97445-v5xqc"] Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.070212 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.559754 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.592065 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d407562c-9b27-4c40-8854-3b62a6178367","Type":"ContainerStarted","Data":"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1"} Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.592106 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d407562c-9b27-4c40-8854-3b62a6178367","Type":"ContainerStarted","Data":"1e8b2830b83f4383adbe4a1e2063dff159d9dd97213617e9736f129a65d75df3"} Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.593656 4741 generic.go:334] "Generic (PLEG): container finished" podID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerID="44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd" exitCode=0 Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.593683 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" event={"ID":"217dfc37-f84b-4114-b2be-b2b55b67ce92","Type":"ContainerDied","Data":"44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd"} Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.593702 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" event={"ID":"217dfc37-f84b-4114-b2be-b2b55b67ce92","Type":"ContainerStarted","Data":"5607c8b39572673129949651751a0231422c938f2dd674f33e22459664db2546"} Sep 29 20:39:28 crc kubenswrapper[4741]: I0929 20:39:28.718369 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:28 crc kubenswrapper[4741]: W0929 20:39:28.722185 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee07c199_ba69_4a3c_a857_f7d646cdf710.slice/crio-f6c420ce72f4b14b897717d2a8f92d47e675a46156b204b7d9d648917944862b WatchSource:0}: Error finding container f6c420ce72f4b14b897717d2a8f92d47e675a46156b204b7d9d648917944862b: Status 404 returned error can't find the container with id f6c420ce72f4b14b897717d2a8f92d47e675a46156b204b7d9d648917944862b Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.605316 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d407562c-9b27-4c40-8854-3b62a6178367","Type":"ContainerStarted","Data":"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0"} Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.605424 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-log" containerID="cri-o://a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1" gracePeriod=30 Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.605500 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-httpd" containerID="cri-o://3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0" gracePeriod=30 Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.607840 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" event={"ID":"217dfc37-f84b-4114-b2be-b2b55b67ce92","Type":"ContainerStarted","Data":"ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a"} Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.608644 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.613088 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee07c199-ba69-4a3c-a857-f7d646cdf710","Type":"ContainerStarted","Data":"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8"} Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.613124 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee07c199-ba69-4a3c-a857-f7d646cdf710","Type":"ContainerStarted","Data":"f6c420ce72f4b14b897717d2a8f92d47e675a46156b204b7d9d648917944862b"} Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.635676 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.6356484030000003 podStartE2EDuration="3.635648403s" podCreationTimestamp="2025-09-29 20:39:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:39:29.633689043 +0000 UTC m=+5411.281478375" watchObservedRunningTime="2025-09-29 20:39:29.635648403 +0000 UTC m=+5411.283437735" Sep 29 20:39:29 crc kubenswrapper[4741]: I0929 20:39:29.656827 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" podStartSLOduration=3.656805828 podStartE2EDuration="3.656805828s" podCreationTimestamp="2025-09-29 20:39:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:39:29.649250894 +0000 UTC m=+5411.297040226" watchObservedRunningTime="2025-09-29 20:39:29.656805828 +0000 UTC m=+5411.304595160" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.206234 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.326711 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-config-data\") pod \"d407562c-9b27-4c40-8854-3b62a6178367\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.326753 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hjpk\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-kube-api-access-5hjpk\") pod \"d407562c-9b27-4c40-8854-3b62a6178367\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.326812 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-scripts\") pod \"d407562c-9b27-4c40-8854-3b62a6178367\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.326844 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-combined-ca-bundle\") pod \"d407562c-9b27-4c40-8854-3b62a6178367\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.326899 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-ceph\") pod \"d407562c-9b27-4c40-8854-3b62a6178367\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.326951 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-httpd-run\") pod \"d407562c-9b27-4c40-8854-3b62a6178367\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.327007 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-logs\") pod \"d407562c-9b27-4c40-8854-3b62a6178367\" (UID: \"d407562c-9b27-4c40-8854-3b62a6178367\") " Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.327587 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-logs" (OuterVolumeSpecName: "logs") pod "d407562c-9b27-4c40-8854-3b62a6178367" (UID: "d407562c-9b27-4c40-8854-3b62a6178367"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.327654 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d407562c-9b27-4c40-8854-3b62a6178367" (UID: "d407562c-9b27-4c40-8854-3b62a6178367"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.328221 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.328308 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d407562c-9b27-4c40-8854-3b62a6178367-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.335233 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-scripts" (OuterVolumeSpecName: "scripts") pod "d407562c-9b27-4c40-8854-3b62a6178367" (UID: "d407562c-9b27-4c40-8854-3b62a6178367"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.335328 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-kube-api-access-5hjpk" (OuterVolumeSpecName: "kube-api-access-5hjpk") pod "d407562c-9b27-4c40-8854-3b62a6178367" (UID: "d407562c-9b27-4c40-8854-3b62a6178367"). InnerVolumeSpecName "kube-api-access-5hjpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.335804 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-ceph" (OuterVolumeSpecName: "ceph") pod "d407562c-9b27-4c40-8854-3b62a6178367" (UID: "d407562c-9b27-4c40-8854-3b62a6178367"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.358038 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d407562c-9b27-4c40-8854-3b62a6178367" (UID: "d407562c-9b27-4c40-8854-3b62a6178367"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.371992 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-config-data" (OuterVolumeSpecName: "config-data") pod "d407562c-9b27-4c40-8854-3b62a6178367" (UID: "d407562c-9b27-4c40-8854-3b62a6178367"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.429900 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.430111 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hjpk\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-kube-api-access-5hjpk\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.430170 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.430270 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d407562c-9b27-4c40-8854-3b62a6178367-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.430356 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d407562c-9b27-4c40-8854-3b62a6178367-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.623676 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee07c199-ba69-4a3c-a857-f7d646cdf710","Type":"ContainerStarted","Data":"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13"} Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.626803 4741 generic.go:334] "Generic (PLEG): container finished" podID="d407562c-9b27-4c40-8854-3b62a6178367" containerID="3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0" exitCode=0 Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.626924 4741 generic.go:334] "Generic (PLEG): container finished" podID="d407562c-9b27-4c40-8854-3b62a6178367" containerID="a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1" exitCode=143 Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.627886 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.637157 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d407562c-9b27-4c40-8854-3b62a6178367","Type":"ContainerDied","Data":"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0"} Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.637309 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d407562c-9b27-4c40-8854-3b62a6178367","Type":"ContainerDied","Data":"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1"} Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.637426 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d407562c-9b27-4c40-8854-3b62a6178367","Type":"ContainerDied","Data":"1e8b2830b83f4383adbe4a1e2063dff159d9dd97213617e9736f129a65d75df3"} Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.637518 4741 scope.go:117] "RemoveContainer" containerID="3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.663496 4741 scope.go:117] "RemoveContainer" containerID="a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.663850 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.663753047 podStartE2EDuration="3.663753047s" podCreationTimestamp="2025-09-29 20:39:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:39:30.657559605 +0000 UTC m=+5412.305348947" watchObservedRunningTime="2025-09-29 20:39:30.663753047 +0000 UTC m=+5412.311542379" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.688929 4741 scope.go:117] "RemoveContainer" containerID="3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0" Sep 29 20:39:30 crc kubenswrapper[4741]: E0929 20:39:30.697592 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0\": container with ID starting with 3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0 not found: ID does not exist" containerID="3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.697722 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0"} err="failed to get container status \"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0\": rpc error: code = NotFound desc = could not find container \"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0\": container with ID starting with 3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0 not found: ID does not exist" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.697759 4741 scope.go:117] "RemoveContainer" containerID="a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1" Sep 29 20:39:30 crc kubenswrapper[4741]: E0929 20:39:30.700703 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1\": container with ID starting with a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1 not found: ID does not exist" containerID="a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.700766 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1"} err="failed to get container status \"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1\": rpc error: code = NotFound desc = could not find container \"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1\": container with ID starting with a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1 not found: ID does not exist" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.700797 4741 scope.go:117] "RemoveContainer" containerID="3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.701162 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0"} err="failed to get container status \"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0\": rpc error: code = NotFound desc = could not find container \"3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0\": container with ID starting with 3f314f5ab021197fab256934210dc85bf080f1bfbf58a0d777eac18fd72323c0 not found: ID does not exist" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.701180 4741 scope.go:117] "RemoveContainer" containerID="a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.702589 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1"} err="failed to get container status \"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1\": rpc error: code = NotFound desc = could not find container \"a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1\": container with ID starting with a0a78aa65428c3983483c33d8bcebe6be3710246faf6eaf9c712df9835b438b1 not found: ID does not exist" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.714279 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.722464 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.731374 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:30 crc kubenswrapper[4741]: E0929 20:39:30.731834 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-httpd" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.731852 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-httpd" Sep 29 20:39:30 crc kubenswrapper[4741]: E0929 20:39:30.731902 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-log" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.731909 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-log" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.732063 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-httpd" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.732080 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d407562c-9b27-4c40-8854-3b62a6178367" containerName="glance-log" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.733243 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.735086 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.738461 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.846431 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.846531 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.846570 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnrr4\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-kube-api-access-gnrr4\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.846588 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-ceph\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.846611 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-config-data\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.846641 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-scripts\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.846700 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-logs\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.948129 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.948171 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnrr4\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-kube-api-access-gnrr4\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.948231 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-ceph\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.948293 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-config-data\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.948311 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-scripts\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.948869 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-logs\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.948958 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.949414 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-logs\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.949480 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.951789 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-scripts\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.951999 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-ceph\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.952316 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-config-data\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.952748 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:30 crc kubenswrapper[4741]: I0929 20:39:30.963694 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnrr4\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-kube-api-access-gnrr4\") pod \"glance-default-external-api-0\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " pod="openstack/glance-default-external-api-0" Sep 29 20:39:31 crc kubenswrapper[4741]: I0929 20:39:31.055590 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:39:31 crc kubenswrapper[4741]: I0929 20:39:31.080991 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:31 crc kubenswrapper[4741]: I0929 20:39:31.107014 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d407562c-9b27-4c40-8854-3b62a6178367" path="/var/lib/kubelet/pods/d407562c-9b27-4c40-8854-3b62a6178367/volumes" Sep 29 20:39:31 crc kubenswrapper[4741]: I0929 20:39:31.648846 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:39:32 crc kubenswrapper[4741]: I0929 20:39:32.649057 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cbf9f732-7174-4fd9-be92-acd5d3571682","Type":"ContainerStarted","Data":"1f348a03ee997223baf508b479abe0f13bfb9f2c307568a6984cda082101b4b6"} Sep 29 20:39:32 crc kubenswrapper[4741]: I0929 20:39:32.649439 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cbf9f732-7174-4fd9-be92-acd5d3571682","Type":"ContainerStarted","Data":"dc501c03ac22dc0c726eaa49e2b647d24785bbcebab5cc11a0abb939a7413a33"} Sep 29 20:39:32 crc kubenswrapper[4741]: I0929 20:39:32.649454 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cbf9f732-7174-4fd9-be92-acd5d3571682","Type":"ContainerStarted","Data":"58a38e3d293df93a9d5f2ee7002c03aeaad4eac75da5fc1a84ee09c15a40f433"} Sep 29 20:39:32 crc kubenswrapper[4741]: I0929 20:39:32.649533 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-log" containerID="cri-o://c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8" gracePeriod=30 Sep 29 20:39:32 crc kubenswrapper[4741]: I0929 20:39:32.649544 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-httpd" containerID="cri-o://ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13" gracePeriod=30 Sep 29 20:39:32 crc kubenswrapper[4741]: I0929 20:39:32.675077 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.675054696 podStartE2EDuration="2.675054696s" podCreationTimestamp="2025-09-29 20:39:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:39:32.668432931 +0000 UTC m=+5414.316222263" watchObservedRunningTime="2025-09-29 20:39:32.675054696 +0000 UTC m=+5414.322844028" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.292171 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.392650 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmqbb\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-kube-api-access-wmqbb\") pod \"ee07c199-ba69-4a3c-a857-f7d646cdf710\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.392712 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-scripts\") pod \"ee07c199-ba69-4a3c-a857-f7d646cdf710\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.392840 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-httpd-run\") pod \"ee07c199-ba69-4a3c-a857-f7d646cdf710\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.392859 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-combined-ca-bundle\") pod \"ee07c199-ba69-4a3c-a857-f7d646cdf710\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.392886 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-ceph\") pod \"ee07c199-ba69-4a3c-a857-f7d646cdf710\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.392913 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-logs\") pod \"ee07c199-ba69-4a3c-a857-f7d646cdf710\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.392944 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-config-data\") pod \"ee07c199-ba69-4a3c-a857-f7d646cdf710\" (UID: \"ee07c199-ba69-4a3c-a857-f7d646cdf710\") " Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.393624 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ee07c199-ba69-4a3c-a857-f7d646cdf710" (UID: "ee07c199-ba69-4a3c-a857-f7d646cdf710"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.393881 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-logs" (OuterVolumeSpecName: "logs") pod "ee07c199-ba69-4a3c-a857-f7d646cdf710" (UID: "ee07c199-ba69-4a3c-a857-f7d646cdf710"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.398853 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-scripts" (OuterVolumeSpecName: "scripts") pod "ee07c199-ba69-4a3c-a857-f7d646cdf710" (UID: "ee07c199-ba69-4a3c-a857-f7d646cdf710"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.399568 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-kube-api-access-wmqbb" (OuterVolumeSpecName: "kube-api-access-wmqbb") pod "ee07c199-ba69-4a3c-a857-f7d646cdf710" (UID: "ee07c199-ba69-4a3c-a857-f7d646cdf710"). InnerVolumeSpecName "kube-api-access-wmqbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.400285 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-ceph" (OuterVolumeSpecName: "ceph") pod "ee07c199-ba69-4a3c-a857-f7d646cdf710" (UID: "ee07c199-ba69-4a3c-a857-f7d646cdf710"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.423688 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee07c199-ba69-4a3c-a857-f7d646cdf710" (UID: "ee07c199-ba69-4a3c-a857-f7d646cdf710"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.445273 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-config-data" (OuterVolumeSpecName: "config-data") pod "ee07c199-ba69-4a3c-a857-f7d646cdf710" (UID: "ee07c199-ba69-4a3c-a857-f7d646cdf710"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.494974 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.495024 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.495038 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.495056 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmqbb\" (UniqueName: \"kubernetes.io/projected/ee07c199-ba69-4a3c-a857-f7d646cdf710-kube-api-access-wmqbb\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.495069 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.495083 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee07c199-ba69-4a3c-a857-f7d646cdf710-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.495096 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee07c199-ba69-4a3c-a857-f7d646cdf710-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.657855 4741 generic.go:334] "Generic (PLEG): container finished" podID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerID="ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13" exitCode=0 Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.657883 4741 generic.go:334] "Generic (PLEG): container finished" podID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerID="c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8" exitCode=143 Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.657940 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.657942 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee07c199-ba69-4a3c-a857-f7d646cdf710","Type":"ContainerDied","Data":"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13"} Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.657994 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee07c199-ba69-4a3c-a857-f7d646cdf710","Type":"ContainerDied","Data":"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8"} Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.658008 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee07c199-ba69-4a3c-a857-f7d646cdf710","Type":"ContainerDied","Data":"f6c420ce72f4b14b897717d2a8f92d47e675a46156b204b7d9d648917944862b"} Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.658025 4741 scope.go:117] "RemoveContainer" containerID="ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.695138 4741 scope.go:117] "RemoveContainer" containerID="c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.720567 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.730648 4741 scope.go:117] "RemoveContainer" containerID="ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13" Sep 29 20:39:33 crc kubenswrapper[4741]: E0929 20:39:33.731039 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13\": container with ID starting with ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13 not found: ID does not exist" containerID="ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.731081 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13"} err="failed to get container status \"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13\": rpc error: code = NotFound desc = could not find container \"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13\": container with ID starting with ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13 not found: ID does not exist" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.731109 4741 scope.go:117] "RemoveContainer" containerID="c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8" Sep 29 20:39:33 crc kubenswrapper[4741]: E0929 20:39:33.731350 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8\": container with ID starting with c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8 not found: ID does not exist" containerID="c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.731414 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8"} err="failed to get container status \"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8\": rpc error: code = NotFound desc = could not find container \"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8\": container with ID starting with c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8 not found: ID does not exist" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.731434 4741 scope.go:117] "RemoveContainer" containerID="ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.731633 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13"} err="failed to get container status \"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13\": rpc error: code = NotFound desc = could not find container \"ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13\": container with ID starting with ce372728dbc920c61958a9175d7b53233c6b6dd257e6d6ed12ab6c7304cf0f13 not found: ID does not exist" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.731653 4741 scope.go:117] "RemoveContainer" containerID="c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.731839 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8"} err="failed to get container status \"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8\": rpc error: code = NotFound desc = could not find container \"c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8\": container with ID starting with c458993fa5936496e9149c0b006847afb3427c88c0201bbf03914c729f5ef2a8 not found: ID does not exist" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.741069 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.752668 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:33 crc kubenswrapper[4741]: E0929 20:39:33.753137 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-httpd" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.753158 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-httpd" Sep 29 20:39:33 crc kubenswrapper[4741]: E0929 20:39:33.753182 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-log" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.753191 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-log" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.753449 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-httpd" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.753470 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" containerName="glance-log" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.754740 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.755310 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.757111 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.799734 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64h9v\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-kube-api-access-64h9v\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.799791 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.799821 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.799873 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-logs\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.799934 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.799956 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.800086 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.903417 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.903496 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64h9v\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-kube-api-access-64h9v\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.903527 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.903561 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.903620 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-logs\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.903681 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.903712 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.917226 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.917563 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-logs\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.917795 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.925121 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.925424 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.928087 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:33 crc kubenswrapper[4741]: I0929 20:39:33.951340 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64h9v\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-kube-api-access-64h9v\") pod \"glance-default-internal-api-0\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:39:34 crc kubenswrapper[4741]: I0929 20:39:34.073457 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:34 crc kubenswrapper[4741]: I0929 20:39:34.595681 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:39:34 crc kubenswrapper[4741]: W0929 20:39:34.595741 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2d875a0_8d33_4254_ad10_0d4d40edfcf9.slice/crio-8d0a22297d00c3f5e6d7ca966f2c22c1a18eabca53d66a5e0308631222f3c224 WatchSource:0}: Error finding container 8d0a22297d00c3f5e6d7ca966f2c22c1a18eabca53d66a5e0308631222f3c224: Status 404 returned error can't find the container with id 8d0a22297d00c3f5e6d7ca966f2c22c1a18eabca53d66a5e0308631222f3c224 Sep 29 20:39:34 crc kubenswrapper[4741]: I0929 20:39:34.680044 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d875a0-8d33-4254-ad10-0d4d40edfcf9","Type":"ContainerStarted","Data":"8d0a22297d00c3f5e6d7ca966f2c22c1a18eabca53d66a5e0308631222f3c224"} Sep 29 20:39:35 crc kubenswrapper[4741]: I0929 20:39:35.089733 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:39:35 crc kubenswrapper[4741]: E0929 20:39:35.090084 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:39:35 crc kubenswrapper[4741]: I0929 20:39:35.097547 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee07c199-ba69-4a3c-a857-f7d646cdf710" path="/var/lib/kubelet/pods/ee07c199-ba69-4a3c-a857-f7d646cdf710/volumes" Sep 29 20:39:35 crc kubenswrapper[4741]: I0929 20:39:35.691361 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d875a0-8d33-4254-ad10-0d4d40edfcf9","Type":"ContainerStarted","Data":"d85377e8f9ab6570cccf4bf2331260524ea0849da2b65cef38d22408a1f59255"} Sep 29 20:39:35 crc kubenswrapper[4741]: I0929 20:39:35.691720 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d875a0-8d33-4254-ad10-0d4d40edfcf9","Type":"ContainerStarted","Data":"0af233296b6dbd422d326bf56f0fc7ae5b97f3738b5678eece65c80e7a88d1aa"} Sep 29 20:39:36 crc kubenswrapper[4741]: I0929 20:39:36.718974 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.718950002 podStartE2EDuration="3.718950002s" podCreationTimestamp="2025-09-29 20:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:39:36.71210034 +0000 UTC m=+5418.359889702" watchObservedRunningTime="2025-09-29 20:39:36.718950002 +0000 UTC m=+5418.366739334" Sep 29 20:39:37 crc kubenswrapper[4741]: I0929 20:39:37.369611 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:39:37 crc kubenswrapper[4741]: I0929 20:39:37.440975 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8677bf69bf-c28js"] Sep 29 20:39:37 crc kubenswrapper[4741]: I0929 20:39:37.441191 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" podUID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerName="dnsmasq-dns" containerID="cri-o://835cc77c0151c98a90f4e774cd7c9e02b4afdba9bfa9bdb89fbf49e24d301acd" gracePeriod=10 Sep 29 20:39:37 crc kubenswrapper[4741]: I0929 20:39:37.727990 4741 generic.go:334] "Generic (PLEG): container finished" podID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerID="835cc77c0151c98a90f4e774cd7c9e02b4afdba9bfa9bdb89fbf49e24d301acd" exitCode=0 Sep 29 20:39:37 crc kubenswrapper[4741]: I0929 20:39:37.728267 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" event={"ID":"0a3a22ac-699b-492e-85ad-3fb34c5697fa","Type":"ContainerDied","Data":"835cc77c0151c98a90f4e774cd7c9e02b4afdba9bfa9bdb89fbf49e24d301acd"} Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.427165 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.486511 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-nb\") pod \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.486672 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl8x2\" (UniqueName: \"kubernetes.io/projected/0a3a22ac-699b-492e-85ad-3fb34c5697fa-kube-api-access-tl8x2\") pod \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.486721 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-dns-svc\") pod \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.486793 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-config\") pod \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.487249 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-sb\") pod \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\" (UID: \"0a3a22ac-699b-492e-85ad-3fb34c5697fa\") " Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.492008 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a3a22ac-699b-492e-85ad-3fb34c5697fa-kube-api-access-tl8x2" (OuterVolumeSpecName: "kube-api-access-tl8x2") pod "0a3a22ac-699b-492e-85ad-3fb34c5697fa" (UID: "0a3a22ac-699b-492e-85ad-3fb34c5697fa"). InnerVolumeSpecName "kube-api-access-tl8x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.526090 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a3a22ac-699b-492e-85ad-3fb34c5697fa" (UID: "0a3a22ac-699b-492e-85ad-3fb34c5697fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.535699 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-config" (OuterVolumeSpecName: "config") pod "0a3a22ac-699b-492e-85ad-3fb34c5697fa" (UID: "0a3a22ac-699b-492e-85ad-3fb34c5697fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.546079 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a3a22ac-699b-492e-85ad-3fb34c5697fa" (UID: "0a3a22ac-699b-492e-85ad-3fb34c5697fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.548471 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a3a22ac-699b-492e-85ad-3fb34c5697fa" (UID: "0a3a22ac-699b-492e-85ad-3fb34c5697fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.588847 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl8x2\" (UniqueName: \"kubernetes.io/projected/0a3a22ac-699b-492e-85ad-3fb34c5697fa-kube-api-access-tl8x2\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.588896 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.588905 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.588913 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.588921 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a3a22ac-699b-492e-85ad-3fb34c5697fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.745741 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" event={"ID":"0a3a22ac-699b-492e-85ad-3fb34c5697fa","Type":"ContainerDied","Data":"bf61d0c79230326598e99e167891b4d7bbc40833995f4e24cc8e76e915e0aeff"} Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.745808 4741 scope.go:117] "RemoveContainer" containerID="835cc77c0151c98a90f4e774cd7c9e02b4afdba9bfa9bdb89fbf49e24d301acd" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.745842 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8677bf69bf-c28js" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.778170 4741 scope.go:117] "RemoveContainer" containerID="c8998e3bdbea1361f89ecb91adf4d07d8f21ea00dbc3b3c5127a283f0632779d" Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.792542 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8677bf69bf-c28js"] Sep 29 20:39:38 crc kubenswrapper[4741]: I0929 20:39:38.799559 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8677bf69bf-c28js"] Sep 29 20:39:39 crc kubenswrapper[4741]: I0929 20:39:39.102666 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" path="/var/lib/kubelet/pods/0a3a22ac-699b-492e-85ad-3fb34c5697fa/volumes" Sep 29 20:39:41 crc kubenswrapper[4741]: I0929 20:39:41.056833 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 20:39:41 crc kubenswrapper[4741]: I0929 20:39:41.057645 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 20:39:41 crc kubenswrapper[4741]: I0929 20:39:41.098722 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 20:39:41 crc kubenswrapper[4741]: I0929 20:39:41.108030 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 20:39:41 crc kubenswrapper[4741]: I0929 20:39:41.771743 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 20:39:41 crc kubenswrapper[4741]: I0929 20:39:41.771805 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 20:39:43 crc kubenswrapper[4741]: I0929 20:39:43.679132 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 20:39:43 crc kubenswrapper[4741]: I0929 20:39:43.702006 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 20:39:44 crc kubenswrapper[4741]: I0929 20:39:44.074534 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:44 crc kubenswrapper[4741]: I0929 20:39:44.074885 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:44 crc kubenswrapper[4741]: I0929 20:39:44.109244 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:44 crc kubenswrapper[4741]: I0929 20:39:44.113877 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:44 crc kubenswrapper[4741]: I0929 20:39:44.809591 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:44 crc kubenswrapper[4741]: I0929 20:39:44.809632 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:46 crc kubenswrapper[4741]: I0929 20:39:46.694105 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:46 crc kubenswrapper[4741]: I0929 20:39:46.740020 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 20:39:47 crc kubenswrapper[4741]: I0929 20:39:47.085900 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:39:47 crc kubenswrapper[4741]: E0929 20:39:47.086381 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.827871 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-4n6h7"] Sep 29 20:39:53 crc kubenswrapper[4741]: E0929 20:39:53.828868 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerName="init" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.828885 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerName="init" Sep 29 20:39:53 crc kubenswrapper[4741]: E0929 20:39:53.828895 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerName="dnsmasq-dns" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.828903 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerName="dnsmasq-dns" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.829144 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a3a22ac-699b-492e-85ad-3fb34c5697fa" containerName="dnsmasq-dns" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.829974 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4n6h7" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.836752 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4n6h7"] Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.849263 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5q89\" (UniqueName: \"kubernetes.io/projected/922b6d69-6d18-4a0c-b81a-e2aa6d103c2f-kube-api-access-d5q89\") pod \"placement-db-create-4n6h7\" (UID: \"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f\") " pod="openstack/placement-db-create-4n6h7" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.951238 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5q89\" (UniqueName: \"kubernetes.io/projected/922b6d69-6d18-4a0c-b81a-e2aa6d103c2f-kube-api-access-d5q89\") pod \"placement-db-create-4n6h7\" (UID: \"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f\") " pod="openstack/placement-db-create-4n6h7" Sep 29 20:39:53 crc kubenswrapper[4741]: I0929 20:39:53.969077 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5q89\" (UniqueName: \"kubernetes.io/projected/922b6d69-6d18-4a0c-b81a-e2aa6d103c2f-kube-api-access-d5q89\") pod \"placement-db-create-4n6h7\" (UID: \"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f\") " pod="openstack/placement-db-create-4n6h7" Sep 29 20:39:54 crc kubenswrapper[4741]: I0929 20:39:54.148107 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4n6h7" Sep 29 20:39:54 crc kubenswrapper[4741]: I0929 20:39:54.577637 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4n6h7"] Sep 29 20:39:54 crc kubenswrapper[4741]: I0929 20:39:54.888731 4741 generic.go:334] "Generic (PLEG): container finished" podID="922b6d69-6d18-4a0c-b81a-e2aa6d103c2f" containerID="e62edaf61dadabb033b60c92a8ef4dcbc4e9791d8b47a597eeb00714c7a2328a" exitCode=0 Sep 29 20:39:54 crc kubenswrapper[4741]: I0929 20:39:54.888774 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4n6h7" event={"ID":"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f","Type":"ContainerDied","Data":"e62edaf61dadabb033b60c92a8ef4dcbc4e9791d8b47a597eeb00714c7a2328a"} Sep 29 20:39:54 crc kubenswrapper[4741]: I0929 20:39:54.888802 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4n6h7" event={"ID":"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f","Type":"ContainerStarted","Data":"fc13a34509736bfec55f1ce953c31fe7f2637fab0775bf0a0a270e2297c614c0"} Sep 29 20:39:56 crc kubenswrapper[4741]: I0929 20:39:56.263811 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4n6h7" Sep 29 20:39:56 crc kubenswrapper[4741]: I0929 20:39:56.389352 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5q89\" (UniqueName: \"kubernetes.io/projected/922b6d69-6d18-4a0c-b81a-e2aa6d103c2f-kube-api-access-d5q89\") pod \"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f\" (UID: \"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f\") " Sep 29 20:39:56 crc kubenswrapper[4741]: I0929 20:39:56.396966 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/922b6d69-6d18-4a0c-b81a-e2aa6d103c2f-kube-api-access-d5q89" (OuterVolumeSpecName: "kube-api-access-d5q89") pod "922b6d69-6d18-4a0c-b81a-e2aa6d103c2f" (UID: "922b6d69-6d18-4a0c-b81a-e2aa6d103c2f"). InnerVolumeSpecName "kube-api-access-d5q89". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:39:56 crc kubenswrapper[4741]: I0929 20:39:56.491979 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5q89\" (UniqueName: \"kubernetes.io/projected/922b6d69-6d18-4a0c-b81a-e2aa6d103c2f-kube-api-access-d5q89\") on node \"crc\" DevicePath \"\"" Sep 29 20:39:56 crc kubenswrapper[4741]: I0929 20:39:56.906719 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4n6h7" event={"ID":"922b6d69-6d18-4a0c-b81a-e2aa6d103c2f","Type":"ContainerDied","Data":"fc13a34509736bfec55f1ce953c31fe7f2637fab0775bf0a0a270e2297c614c0"} Sep 29 20:39:56 crc kubenswrapper[4741]: I0929 20:39:56.906760 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc13a34509736bfec55f1ce953c31fe7f2637fab0775bf0a0a270e2297c614c0" Sep 29 20:39:56 crc kubenswrapper[4741]: I0929 20:39:56.906802 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4n6h7" Sep 29 20:40:01 crc kubenswrapper[4741]: I0929 20:40:01.085983 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:40:01 crc kubenswrapper[4741]: E0929 20:40:01.087038 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:40:03 crc kubenswrapper[4741]: I0929 20:40:03.951772 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a214-account-create-jlwc9"] Sep 29 20:40:03 crc kubenswrapper[4741]: E0929 20:40:03.952430 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922b6d69-6d18-4a0c-b81a-e2aa6d103c2f" containerName="mariadb-database-create" Sep 29 20:40:03 crc kubenswrapper[4741]: I0929 20:40:03.952443 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="922b6d69-6d18-4a0c-b81a-e2aa6d103c2f" containerName="mariadb-database-create" Sep 29 20:40:03 crc kubenswrapper[4741]: I0929 20:40:03.952656 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="922b6d69-6d18-4a0c-b81a-e2aa6d103c2f" containerName="mariadb-database-create" Sep 29 20:40:03 crc kubenswrapper[4741]: I0929 20:40:03.953244 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a214-account-create-jlwc9" Sep 29 20:40:03 crc kubenswrapper[4741]: I0929 20:40:03.956333 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Sep 29 20:40:03 crc kubenswrapper[4741]: I0929 20:40:03.968910 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a214-account-create-jlwc9"] Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.132704 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-246sb\" (UniqueName: \"kubernetes.io/projected/5007e794-725b-48f2-916b-8e412e4780fc-kube-api-access-246sb\") pod \"placement-a214-account-create-jlwc9\" (UID: \"5007e794-725b-48f2-916b-8e412e4780fc\") " pod="openstack/placement-a214-account-create-jlwc9" Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.233829 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-246sb\" (UniqueName: \"kubernetes.io/projected/5007e794-725b-48f2-916b-8e412e4780fc-kube-api-access-246sb\") pod \"placement-a214-account-create-jlwc9\" (UID: \"5007e794-725b-48f2-916b-8e412e4780fc\") " pod="openstack/placement-a214-account-create-jlwc9" Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.253190 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-246sb\" (UniqueName: \"kubernetes.io/projected/5007e794-725b-48f2-916b-8e412e4780fc-kube-api-access-246sb\") pod \"placement-a214-account-create-jlwc9\" (UID: \"5007e794-725b-48f2-916b-8e412e4780fc\") " pod="openstack/placement-a214-account-create-jlwc9" Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.277239 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a214-account-create-jlwc9" Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.703239 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a214-account-create-jlwc9"] Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.988688 4741 generic.go:334] "Generic (PLEG): container finished" podID="5007e794-725b-48f2-916b-8e412e4780fc" containerID="26e8685f0b49214da518ce90beb6aa2904ce8226b2bb7c38cd7954a956920cef" exitCode=0 Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.988739 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a214-account-create-jlwc9" event={"ID":"5007e794-725b-48f2-916b-8e412e4780fc","Type":"ContainerDied","Data":"26e8685f0b49214da518ce90beb6aa2904ce8226b2bb7c38cd7954a956920cef"} Sep 29 20:40:04 crc kubenswrapper[4741]: I0929 20:40:04.989068 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a214-account-create-jlwc9" event={"ID":"5007e794-725b-48f2-916b-8e412e4780fc","Type":"ContainerStarted","Data":"c86cbfd0eba52cbcde209a5d6cedefe0ea8e89ef4c49d81d95e8b11a8405e601"} Sep 29 20:40:06 crc kubenswrapper[4741]: I0929 20:40:06.384265 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a214-account-create-jlwc9" Sep 29 20:40:06 crc kubenswrapper[4741]: I0929 20:40:06.473583 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-246sb\" (UniqueName: \"kubernetes.io/projected/5007e794-725b-48f2-916b-8e412e4780fc-kube-api-access-246sb\") pod \"5007e794-725b-48f2-916b-8e412e4780fc\" (UID: \"5007e794-725b-48f2-916b-8e412e4780fc\") " Sep 29 20:40:06 crc kubenswrapper[4741]: I0929 20:40:06.481216 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5007e794-725b-48f2-916b-8e412e4780fc-kube-api-access-246sb" (OuterVolumeSpecName: "kube-api-access-246sb") pod "5007e794-725b-48f2-916b-8e412e4780fc" (UID: "5007e794-725b-48f2-916b-8e412e4780fc"). InnerVolumeSpecName "kube-api-access-246sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:40:06 crc kubenswrapper[4741]: I0929 20:40:06.575556 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-246sb\" (UniqueName: \"kubernetes.io/projected/5007e794-725b-48f2-916b-8e412e4780fc-kube-api-access-246sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:07 crc kubenswrapper[4741]: I0929 20:40:07.010329 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a214-account-create-jlwc9" event={"ID":"5007e794-725b-48f2-916b-8e412e4780fc","Type":"ContainerDied","Data":"c86cbfd0eba52cbcde209a5d6cedefe0ea8e89ef4c49d81d95e8b11a8405e601"} Sep 29 20:40:07 crc kubenswrapper[4741]: I0929 20:40:07.010366 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c86cbfd0eba52cbcde209a5d6cedefe0ea8e89ef4c49d81d95e8b11a8405e601" Sep 29 20:40:07 crc kubenswrapper[4741]: I0929 20:40:07.010370 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a214-account-create-jlwc9" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.248979 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dffb944f7-sx7bc"] Sep 29 20:40:09 crc kubenswrapper[4741]: E0929 20:40:09.250971 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5007e794-725b-48f2-916b-8e412e4780fc" containerName="mariadb-account-create" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.251011 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5007e794-725b-48f2-916b-8e412e4780fc" containerName="mariadb-account-create" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.251264 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5007e794-725b-48f2-916b-8e412e4780fc" containerName="mariadb-account-create" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.252640 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.275945 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dffb944f7-sx7bc"] Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.295441 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nwzl6"] Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.296458 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.298873 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-psnkk" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.298921 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.299885 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.304543 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nwzl6"] Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.422348 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-sb\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.422439 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a716d83-536e-41ad-b8df-b55cfbfac7a1-logs\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.422570 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8hbk\" (UniqueName: \"kubernetes.io/projected/d8f88100-58cc-4534-b686-1656b22af01f-kube-api-access-m8hbk\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.422622 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-scripts\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.422670 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v79r6\" (UniqueName: \"kubernetes.io/projected/7a716d83-536e-41ad-b8df-b55cfbfac7a1-kube-api-access-v79r6\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.422762 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-config-data\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.422896 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-nb\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.423067 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-dns-svc\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.423124 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-combined-ca-bundle\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.423187 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-config\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.524761 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-combined-ca-bundle\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525662 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-config\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525752 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-sb\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525785 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a716d83-536e-41ad-b8df-b55cfbfac7a1-logs\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525832 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8hbk\" (UniqueName: \"kubernetes.io/projected/d8f88100-58cc-4534-b686-1656b22af01f-kube-api-access-m8hbk\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525862 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-scripts\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525916 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v79r6\" (UniqueName: \"kubernetes.io/projected/7a716d83-536e-41ad-b8df-b55cfbfac7a1-kube-api-access-v79r6\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525953 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-config-data\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.525989 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-nb\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.526044 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-dns-svc\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.526340 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a716d83-536e-41ad-b8df-b55cfbfac7a1-logs\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.527100 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-dns-svc\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.527172 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-nb\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.527318 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-config\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.527407 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-sb\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.529751 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-scripts\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.529924 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-combined-ca-bundle\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.530658 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-config-data\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.543542 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8hbk\" (UniqueName: \"kubernetes.io/projected/d8f88100-58cc-4534-b686-1656b22af01f-kube-api-access-m8hbk\") pod \"dnsmasq-dns-dffb944f7-sx7bc\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.544205 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v79r6\" (UniqueName: \"kubernetes.io/projected/7a716d83-536e-41ad-b8df-b55cfbfac7a1-kube-api-access-v79r6\") pod \"placement-db-sync-nwzl6\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.586087 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:09 crc kubenswrapper[4741]: I0929 20:40:09.614806 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:10 crc kubenswrapper[4741]: I0929 20:40:10.080968 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nwzl6"] Sep 29 20:40:10 crc kubenswrapper[4741]: I0929 20:40:10.152471 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dffb944f7-sx7bc"] Sep 29 20:40:10 crc kubenswrapper[4741]: W0929 20:40:10.162545 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8f88100_58cc_4534_b686_1656b22af01f.slice/crio-17ca3735e41a71f4254d1f47bfbcd906f88867ce447180447d7d445f9661601e WatchSource:0}: Error finding container 17ca3735e41a71f4254d1f47bfbcd906f88867ce447180447d7d445f9661601e: Status 404 returned error can't find the container with id 17ca3735e41a71f4254d1f47bfbcd906f88867ce447180447d7d445f9661601e Sep 29 20:40:11 crc kubenswrapper[4741]: I0929 20:40:11.048690 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nwzl6" event={"ID":"7a716d83-536e-41ad-b8df-b55cfbfac7a1","Type":"ContainerStarted","Data":"fef575ca088dbdbd8d09178b957915cfb88901de31de6029240d451ec60c97a1"} Sep 29 20:40:11 crc kubenswrapper[4741]: I0929 20:40:11.049024 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nwzl6" event={"ID":"7a716d83-536e-41ad-b8df-b55cfbfac7a1","Type":"ContainerStarted","Data":"f043c738e1bc6ea92043aa28736adb502017b96df5cedbd323377fde82991bbc"} Sep 29 20:40:11 crc kubenswrapper[4741]: I0929 20:40:11.051526 4741 generic.go:334] "Generic (PLEG): container finished" podID="d8f88100-58cc-4534-b686-1656b22af01f" containerID="16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529" exitCode=0 Sep 29 20:40:11 crc kubenswrapper[4741]: I0929 20:40:11.051578 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" event={"ID":"d8f88100-58cc-4534-b686-1656b22af01f","Type":"ContainerDied","Data":"16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529"} Sep 29 20:40:11 crc kubenswrapper[4741]: I0929 20:40:11.051616 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" event={"ID":"d8f88100-58cc-4534-b686-1656b22af01f","Type":"ContainerStarted","Data":"17ca3735e41a71f4254d1f47bfbcd906f88867ce447180447d7d445f9661601e"} Sep 29 20:40:11 crc kubenswrapper[4741]: I0929 20:40:11.072870 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nwzl6" podStartSLOduration=2.072849371 podStartE2EDuration="2.072849371s" podCreationTimestamp="2025-09-29 20:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:40:11.068015472 +0000 UTC m=+5452.715804874" watchObservedRunningTime="2025-09-29 20:40:11.072849371 +0000 UTC m=+5452.720638703" Sep 29 20:40:12 crc kubenswrapper[4741]: I0929 20:40:12.060250 4741 generic.go:334] "Generic (PLEG): container finished" podID="7a716d83-536e-41ad-b8df-b55cfbfac7a1" containerID="fef575ca088dbdbd8d09178b957915cfb88901de31de6029240d451ec60c97a1" exitCode=0 Sep 29 20:40:12 crc kubenswrapper[4741]: I0929 20:40:12.060323 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nwzl6" event={"ID":"7a716d83-536e-41ad-b8df-b55cfbfac7a1","Type":"ContainerDied","Data":"fef575ca088dbdbd8d09178b957915cfb88901de31de6029240d451ec60c97a1"} Sep 29 20:40:12 crc kubenswrapper[4741]: I0929 20:40:12.062137 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" event={"ID":"d8f88100-58cc-4534-b686-1656b22af01f","Type":"ContainerStarted","Data":"c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620"} Sep 29 20:40:12 crc kubenswrapper[4741]: I0929 20:40:12.062443 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:12 crc kubenswrapper[4741]: I0929 20:40:12.097487 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" podStartSLOduration=3.097472598 podStartE2EDuration="3.097472598s" podCreationTimestamp="2025-09-29 20:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:40:12.092104972 +0000 UTC m=+5453.739894304" watchObservedRunningTime="2025-09-29 20:40:12.097472598 +0000 UTC m=+5453.745261920" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.086533 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:40:13 crc kubenswrapper[4741]: E0929 20:40:13.087268 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.428238 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.527178 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-combined-ca-bundle\") pod \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.527290 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-scripts\") pod \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.527320 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a716d83-536e-41ad-b8df-b55cfbfac7a1-logs\") pod \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.527363 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v79r6\" (UniqueName: \"kubernetes.io/projected/7a716d83-536e-41ad-b8df-b55cfbfac7a1-kube-api-access-v79r6\") pod \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.527436 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-config-data\") pod \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\" (UID: \"7a716d83-536e-41ad-b8df-b55cfbfac7a1\") " Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.527731 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a716d83-536e-41ad-b8df-b55cfbfac7a1-logs" (OuterVolumeSpecName: "logs") pod "7a716d83-536e-41ad-b8df-b55cfbfac7a1" (UID: "7a716d83-536e-41ad-b8df-b55cfbfac7a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.535616 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-scripts" (OuterVolumeSpecName: "scripts") pod "7a716d83-536e-41ad-b8df-b55cfbfac7a1" (UID: "7a716d83-536e-41ad-b8df-b55cfbfac7a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.536553 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a716d83-536e-41ad-b8df-b55cfbfac7a1-kube-api-access-v79r6" (OuterVolumeSpecName: "kube-api-access-v79r6") pod "7a716d83-536e-41ad-b8df-b55cfbfac7a1" (UID: "7a716d83-536e-41ad-b8df-b55cfbfac7a1"). InnerVolumeSpecName "kube-api-access-v79r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.556904 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-config-data" (OuterVolumeSpecName: "config-data") pod "7a716d83-536e-41ad-b8df-b55cfbfac7a1" (UID: "7a716d83-536e-41ad-b8df-b55cfbfac7a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.558688 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a716d83-536e-41ad-b8df-b55cfbfac7a1" (UID: "7a716d83-536e-41ad-b8df-b55cfbfac7a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.630166 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.630206 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a716d83-536e-41ad-b8df-b55cfbfac7a1-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.630419 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v79r6\" (UniqueName: \"kubernetes.io/projected/7a716d83-536e-41ad-b8df-b55cfbfac7a1-kube-api-access-v79r6\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.630436 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:13 crc kubenswrapper[4741]: I0929 20:40:13.630455 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a716d83-536e-41ad-b8df-b55cfbfac7a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.091619 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nwzl6" event={"ID":"7a716d83-536e-41ad-b8df-b55cfbfac7a1","Type":"ContainerDied","Data":"f043c738e1bc6ea92043aa28736adb502017b96df5cedbd323377fde82991bbc"} Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.091653 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f043c738e1bc6ea92043aa28736adb502017b96df5cedbd323377fde82991bbc" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.091687 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nwzl6" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.171586 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-649f4d5b76-lcsk7"] Sep 29 20:40:14 crc kubenswrapper[4741]: E0929 20:40:14.172459 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a716d83-536e-41ad-b8df-b55cfbfac7a1" containerName="placement-db-sync" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.172567 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a716d83-536e-41ad-b8df-b55cfbfac7a1" containerName="placement-db-sync" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.172881 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a716d83-536e-41ad-b8df-b55cfbfac7a1" containerName="placement-db-sync" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.174124 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.175706 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-649f4d5b76-lcsk7"] Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.180357 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-psnkk" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.180579 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.180694 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.340878 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8cbl\" (UniqueName: \"kubernetes.io/projected/3233508f-ad82-4a1e-aa95-d647a3fdff0b-kube-api-access-w8cbl\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.341000 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-combined-ca-bundle\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.341093 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3233508f-ad82-4a1e-aa95-d647a3fdff0b-logs\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.341249 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-config-data\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.341323 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-scripts\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.443912 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-config-data\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.443990 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-scripts\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.444080 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8cbl\" (UniqueName: \"kubernetes.io/projected/3233508f-ad82-4a1e-aa95-d647a3fdff0b-kube-api-access-w8cbl\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.444127 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-combined-ca-bundle\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.444166 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3233508f-ad82-4a1e-aa95-d647a3fdff0b-logs\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.444625 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3233508f-ad82-4a1e-aa95-d647a3fdff0b-logs\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.448869 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-combined-ca-bundle\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.449033 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-config-data\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.454669 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3233508f-ad82-4a1e-aa95-d647a3fdff0b-scripts\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.461617 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8cbl\" (UniqueName: \"kubernetes.io/projected/3233508f-ad82-4a1e-aa95-d647a3fdff0b-kube-api-access-w8cbl\") pod \"placement-649f4d5b76-lcsk7\" (UID: \"3233508f-ad82-4a1e-aa95-d647a3fdff0b\") " pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.511003 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:14 crc kubenswrapper[4741]: I0929 20:40:14.931432 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-649f4d5b76-lcsk7"] Sep 29 20:40:14 crc kubenswrapper[4741]: W0929 20:40:14.931596 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3233508f_ad82_4a1e_aa95_d647a3fdff0b.slice/crio-942595c398e76c1d463a36e2d13a3c02c66050d17525707cdcbd52ff56f2654b WatchSource:0}: Error finding container 942595c398e76c1d463a36e2d13a3c02c66050d17525707cdcbd52ff56f2654b: Status 404 returned error can't find the container with id 942595c398e76c1d463a36e2d13a3c02c66050d17525707cdcbd52ff56f2654b Sep 29 20:40:15 crc kubenswrapper[4741]: I0929 20:40:15.100037 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649f4d5b76-lcsk7" event={"ID":"3233508f-ad82-4a1e-aa95-d647a3fdff0b","Type":"ContainerStarted","Data":"942595c398e76c1d463a36e2d13a3c02c66050d17525707cdcbd52ff56f2654b"} Sep 29 20:40:16 crc kubenswrapper[4741]: I0929 20:40:16.111776 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649f4d5b76-lcsk7" event={"ID":"3233508f-ad82-4a1e-aa95-d647a3fdff0b","Type":"ContainerStarted","Data":"1c05235c2878ada6612664532179f06d2df93918454596e8ac26100825fabfa2"} Sep 29 20:40:16 crc kubenswrapper[4741]: I0929 20:40:16.112604 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:16 crc kubenswrapper[4741]: I0929 20:40:16.112697 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649f4d5b76-lcsk7" event={"ID":"3233508f-ad82-4a1e-aa95-d647a3fdff0b","Type":"ContainerStarted","Data":"85dafb4926887302c88a04706f29f054b0fdb971ce27989e0cf8b6940220edb8"} Sep 29 20:40:16 crc kubenswrapper[4741]: I0929 20:40:16.139308 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-649f4d5b76-lcsk7" podStartSLOduration=2.139290909 podStartE2EDuration="2.139290909s" podCreationTimestamp="2025-09-29 20:40:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:40:16.13479211 +0000 UTC m=+5457.782581462" watchObservedRunningTime="2025-09-29 20:40:16.139290909 +0000 UTC m=+5457.787080241" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.127737 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.178544 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cthf6"] Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.180566 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.185258 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cthf6"] Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.296019 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-catalog-content\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.296112 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-utilities\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.296212 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bxsx\" (UniqueName: \"kubernetes.io/projected/a2b342dd-ba60-4755-bf40-173560a918ca-kube-api-access-2bxsx\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.398376 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-catalog-content\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.398503 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-utilities\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.398548 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bxsx\" (UniqueName: \"kubernetes.io/projected/a2b342dd-ba60-4755-bf40-173560a918ca-kube-api-access-2bxsx\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.398875 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-catalog-content\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.398892 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-utilities\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.420290 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bxsx\" (UniqueName: \"kubernetes.io/projected/a2b342dd-ba60-4755-bf40-173560a918ca-kube-api-access-2bxsx\") pod \"community-operators-cthf6\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.512721 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:17 crc kubenswrapper[4741]: I0929 20:40:17.977038 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cthf6"] Sep 29 20:40:17 crc kubenswrapper[4741]: W0929 20:40:17.983443 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2b342dd_ba60_4755_bf40_173560a918ca.slice/crio-46faa0bdec05e2558c010c0ec7d7c4ed665672b585cb3529ff0b8f029c612639 WatchSource:0}: Error finding container 46faa0bdec05e2558c010c0ec7d7c4ed665672b585cb3529ff0b8f029c612639: Status 404 returned error can't find the container with id 46faa0bdec05e2558c010c0ec7d7c4ed665672b585cb3529ff0b8f029c612639 Sep 29 20:40:18 crc kubenswrapper[4741]: I0929 20:40:18.136062 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cthf6" event={"ID":"a2b342dd-ba60-4755-bf40-173560a918ca","Type":"ContainerStarted","Data":"46faa0bdec05e2558c010c0ec7d7c4ed665672b585cb3529ff0b8f029c612639"} Sep 29 20:40:19 crc kubenswrapper[4741]: I0929 20:40:19.144522 4741 generic.go:334] "Generic (PLEG): container finished" podID="a2b342dd-ba60-4755-bf40-173560a918ca" containerID="9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08" exitCode=0 Sep 29 20:40:19 crc kubenswrapper[4741]: I0929 20:40:19.144595 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cthf6" event={"ID":"a2b342dd-ba60-4755-bf40-173560a918ca","Type":"ContainerDied","Data":"9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08"} Sep 29 20:40:19 crc kubenswrapper[4741]: I0929 20:40:19.587728 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:40:19 crc kubenswrapper[4741]: I0929 20:40:19.665266 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744dc97445-v5xqc"] Sep 29 20:40:19 crc kubenswrapper[4741]: I0929 20:40:19.665560 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" podUID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerName="dnsmasq-dns" containerID="cri-o://ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a" gracePeriod=10 Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.118782 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.196548 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cthf6" event={"ID":"a2b342dd-ba60-4755-bf40-173560a918ca","Type":"ContainerStarted","Data":"272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92"} Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.207917 4741 generic.go:334] "Generic (PLEG): container finished" podID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerID="ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a" exitCode=0 Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.207961 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" event={"ID":"217dfc37-f84b-4114-b2be-b2b55b67ce92","Type":"ContainerDied","Data":"ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a"} Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.207988 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" event={"ID":"217dfc37-f84b-4114-b2be-b2b55b67ce92","Type":"ContainerDied","Data":"5607c8b39572673129949651751a0231422c938f2dd674f33e22459664db2546"} Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.208006 4741 scope.go:117] "RemoveContainer" containerID="ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.208145 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744dc97445-v5xqc" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.248291 4741 scope.go:117] "RemoveContainer" containerID="44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.249070 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-config\") pod \"217dfc37-f84b-4114-b2be-b2b55b67ce92\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.249155 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-dns-svc\") pod \"217dfc37-f84b-4114-b2be-b2b55b67ce92\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.250209 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-sb\") pod \"217dfc37-f84b-4114-b2be-b2b55b67ce92\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.250272 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-nb\") pod \"217dfc37-f84b-4114-b2be-b2b55b67ce92\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.250411 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7dck\" (UniqueName: \"kubernetes.io/projected/217dfc37-f84b-4114-b2be-b2b55b67ce92-kube-api-access-h7dck\") pod \"217dfc37-f84b-4114-b2be-b2b55b67ce92\" (UID: \"217dfc37-f84b-4114-b2be-b2b55b67ce92\") " Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.260708 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/217dfc37-f84b-4114-b2be-b2b55b67ce92-kube-api-access-h7dck" (OuterVolumeSpecName: "kube-api-access-h7dck") pod "217dfc37-f84b-4114-b2be-b2b55b67ce92" (UID: "217dfc37-f84b-4114-b2be-b2b55b67ce92"). InnerVolumeSpecName "kube-api-access-h7dck". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.275201 4741 scope.go:117] "RemoveContainer" containerID="ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a" Sep 29 20:40:20 crc kubenswrapper[4741]: E0929 20:40:20.276064 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a\": container with ID starting with ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a not found: ID does not exist" containerID="ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.276101 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a"} err="failed to get container status \"ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a\": rpc error: code = NotFound desc = could not find container \"ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a\": container with ID starting with ccd78653ad6ec507fdc42fcf57902b3297fc144510929e20793911e67bb0b49a not found: ID does not exist" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.276122 4741 scope.go:117] "RemoveContainer" containerID="44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd" Sep 29 20:40:20 crc kubenswrapper[4741]: E0929 20:40:20.276548 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd\": container with ID starting with 44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd not found: ID does not exist" containerID="44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.276586 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd"} err="failed to get container status \"44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd\": rpc error: code = NotFound desc = could not find container \"44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd\": container with ID starting with 44cb214a7148f78b95e96c8d1a290293b13a9a68c4a11cc99cda90a698ab08fd not found: ID does not exist" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.302986 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "217dfc37-f84b-4114-b2be-b2b55b67ce92" (UID: "217dfc37-f84b-4114-b2be-b2b55b67ce92"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.304225 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-config" (OuterVolumeSpecName: "config") pod "217dfc37-f84b-4114-b2be-b2b55b67ce92" (UID: "217dfc37-f84b-4114-b2be-b2b55b67ce92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.332808 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "217dfc37-f84b-4114-b2be-b2b55b67ce92" (UID: "217dfc37-f84b-4114-b2be-b2b55b67ce92"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.345833 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "217dfc37-f84b-4114-b2be-b2b55b67ce92" (UID: "217dfc37-f84b-4114-b2be-b2b55b67ce92"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.353059 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.353087 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.353097 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.353108 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7dck\" (UniqueName: \"kubernetes.io/projected/217dfc37-f84b-4114-b2be-b2b55b67ce92-kube-api-access-h7dck\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.353117 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/217dfc37-f84b-4114-b2be-b2b55b67ce92-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.538951 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744dc97445-v5xqc"] Sep 29 20:40:20 crc kubenswrapper[4741]: I0929 20:40:20.546472 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-744dc97445-v5xqc"] Sep 29 20:40:21 crc kubenswrapper[4741]: I0929 20:40:21.104204 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="217dfc37-f84b-4114-b2be-b2b55b67ce92" path="/var/lib/kubelet/pods/217dfc37-f84b-4114-b2be-b2b55b67ce92/volumes" Sep 29 20:40:21 crc kubenswrapper[4741]: I0929 20:40:21.217710 4741 generic.go:334] "Generic (PLEG): container finished" podID="a2b342dd-ba60-4755-bf40-173560a918ca" containerID="272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92" exitCode=0 Sep 29 20:40:21 crc kubenswrapper[4741]: I0929 20:40:21.217769 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cthf6" event={"ID":"a2b342dd-ba60-4755-bf40-173560a918ca","Type":"ContainerDied","Data":"272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92"} Sep 29 20:40:21 crc kubenswrapper[4741]: I0929 20:40:21.217796 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cthf6" event={"ID":"a2b342dd-ba60-4755-bf40-173560a918ca","Type":"ContainerStarted","Data":"93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9"} Sep 29 20:40:21 crc kubenswrapper[4741]: I0929 20:40:21.253433 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cthf6" podStartSLOduration=2.814950328 podStartE2EDuration="4.253383743s" podCreationTimestamp="2025-09-29 20:40:17 +0000 UTC" firstStartedPulling="2025-09-29 20:40:19.146837596 +0000 UTC m=+5460.794626938" lastFinishedPulling="2025-09-29 20:40:20.585271021 +0000 UTC m=+5462.233060353" observedRunningTime="2025-09-29 20:40:21.243069023 +0000 UTC m=+5462.890858355" watchObservedRunningTime="2025-09-29 20:40:21.253383743 +0000 UTC m=+5462.901173115" Sep 29 20:40:24 crc kubenswrapper[4741]: I0929 20:40:24.094649 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:40:24 crc kubenswrapper[4741]: E0929 20:40:24.095202 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:40:27 crc kubenswrapper[4741]: I0929 20:40:27.513728 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:27 crc kubenswrapper[4741]: I0929 20:40:27.514186 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:27 crc kubenswrapper[4741]: I0929 20:40:27.585335 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:28 crc kubenswrapper[4741]: I0929 20:40:28.330463 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:28 crc kubenswrapper[4741]: I0929 20:40:28.940520 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cthf6"] Sep 29 20:40:30 crc kubenswrapper[4741]: I0929 20:40:30.303937 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cthf6" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="registry-server" containerID="cri-o://93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9" gracePeriod=2 Sep 29 20:40:30 crc kubenswrapper[4741]: I0929 20:40:30.894230 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.083383 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bxsx\" (UniqueName: \"kubernetes.io/projected/a2b342dd-ba60-4755-bf40-173560a918ca-kube-api-access-2bxsx\") pod \"a2b342dd-ba60-4755-bf40-173560a918ca\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.083572 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-utilities\") pod \"a2b342dd-ba60-4755-bf40-173560a918ca\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.083683 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-catalog-content\") pod \"a2b342dd-ba60-4755-bf40-173560a918ca\" (UID: \"a2b342dd-ba60-4755-bf40-173560a918ca\") " Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.087171 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-utilities" (OuterVolumeSpecName: "utilities") pod "a2b342dd-ba60-4755-bf40-173560a918ca" (UID: "a2b342dd-ba60-4755-bf40-173560a918ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.090024 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b342dd-ba60-4755-bf40-173560a918ca-kube-api-access-2bxsx" (OuterVolumeSpecName: "kube-api-access-2bxsx") pod "a2b342dd-ba60-4755-bf40-173560a918ca" (UID: "a2b342dd-ba60-4755-bf40-173560a918ca"). InnerVolumeSpecName "kube-api-access-2bxsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.137537 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2b342dd-ba60-4755-bf40-173560a918ca" (UID: "a2b342dd-ba60-4755-bf40-173560a918ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.186110 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bxsx\" (UniqueName: \"kubernetes.io/projected/a2b342dd-ba60-4755-bf40-173560a918ca-kube-api-access-2bxsx\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.186149 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.186162 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b342dd-ba60-4755-bf40-173560a918ca-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.320934 4741 generic.go:334] "Generic (PLEG): container finished" podID="a2b342dd-ba60-4755-bf40-173560a918ca" containerID="93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9" exitCode=0 Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.321045 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cthf6" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.321029 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cthf6" event={"ID":"a2b342dd-ba60-4755-bf40-173560a918ca","Type":"ContainerDied","Data":"93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9"} Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.321525 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cthf6" event={"ID":"a2b342dd-ba60-4755-bf40-173560a918ca","Type":"ContainerDied","Data":"46faa0bdec05e2558c010c0ec7d7c4ed665672b585cb3529ff0b8f029c612639"} Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.321574 4741 scope.go:117] "RemoveContainer" containerID="93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.345943 4741 scope.go:117] "RemoveContainer" containerID="272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.376491 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cthf6"] Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.388443 4741 scope.go:117] "RemoveContainer" containerID="9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.389783 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cthf6"] Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.426054 4741 scope.go:117] "RemoveContainer" containerID="93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9" Sep 29 20:40:31 crc kubenswrapper[4741]: E0929 20:40:31.426787 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9\": container with ID starting with 93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9 not found: ID does not exist" containerID="93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.426926 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9"} err="failed to get container status \"93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9\": rpc error: code = NotFound desc = could not find container \"93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9\": container with ID starting with 93cf15d41c9e9d5813e37282af40349aa8158b50700eb5e7381f7212a9f988b9 not found: ID does not exist" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.426999 4741 scope.go:117] "RemoveContainer" containerID="272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92" Sep 29 20:40:31 crc kubenswrapper[4741]: E0929 20:40:31.427616 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92\": container with ID starting with 272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92 not found: ID does not exist" containerID="272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.427668 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92"} err="failed to get container status \"272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92\": rpc error: code = NotFound desc = could not find container \"272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92\": container with ID starting with 272277ad346c7d33fc5eccea1db7590da5a7065dbc7f0a3486ca8bc20dd38a92 not found: ID does not exist" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.427705 4741 scope.go:117] "RemoveContainer" containerID="9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08" Sep 29 20:40:31 crc kubenswrapper[4741]: E0929 20:40:31.428241 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08\": container with ID starting with 9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08 not found: ID does not exist" containerID="9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08" Sep 29 20:40:31 crc kubenswrapper[4741]: I0929 20:40:31.428268 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08"} err="failed to get container status \"9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08\": rpc error: code = NotFound desc = could not find container \"9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08\": container with ID starting with 9d5e6fbf1fda22d05fb7a18d5df0e2f4b4fba5745270f87e3ba51799d7877c08 not found: ID does not exist" Sep 29 20:40:33 crc kubenswrapper[4741]: I0929 20:40:33.095845 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" path="/var/lib/kubelet/pods/a2b342dd-ba60-4755-bf40-173560a918ca/volumes" Sep 29 20:40:38 crc kubenswrapper[4741]: I0929 20:40:38.086880 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:40:38 crc kubenswrapper[4741]: E0929 20:40:38.087438 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:40:45 crc kubenswrapper[4741]: I0929 20:40:45.494819 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:45 crc kubenswrapper[4741]: I0929 20:40:45.498148 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-649f4d5b76-lcsk7" Sep 29 20:40:52 crc kubenswrapper[4741]: I0929 20:40:52.086085 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:40:52 crc kubenswrapper[4741]: E0929 20:40:52.086881 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.585327 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-98rj6"] Sep 29 20:41:05 crc kubenswrapper[4741]: E0929 20:41:05.586215 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="extract-utilities" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.586233 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="extract-utilities" Sep 29 20:41:05 crc kubenswrapper[4741]: E0929 20:41:05.586244 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerName="dnsmasq-dns" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.586252 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerName="dnsmasq-dns" Sep 29 20:41:05 crc kubenswrapper[4741]: E0929 20:41:05.586271 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerName="init" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.586279 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerName="init" Sep 29 20:41:05 crc kubenswrapper[4741]: E0929 20:41:05.586302 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="registry-server" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.586310 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="registry-server" Sep 29 20:41:05 crc kubenswrapper[4741]: E0929 20:41:05.586340 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="extract-content" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.586348 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="extract-content" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.589647 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b342dd-ba60-4755-bf40-173560a918ca" containerName="registry-server" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.589692 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="217dfc37-f84b-4114-b2be-b2b55b67ce92" containerName="dnsmasq-dns" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.590450 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98rj6" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.608360 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-98rj6"] Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.676325 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-bf56t"] Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.677348 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bf56t" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.687359 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7mcn\" (UniqueName: \"kubernetes.io/projected/9a3dce1e-ade4-4090-a9c5-97238453667f-kube-api-access-k7mcn\") pod \"nova-api-db-create-98rj6\" (UID: \"9a3dce1e-ade4-4090-a9c5-97238453667f\") " pod="openstack/nova-api-db-create-98rj6" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.687697 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bf56t"] Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.775767 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-djpqv"] Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.776982 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-djpqv" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.787299 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-djpqv"] Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.789196 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6vbq\" (UniqueName: \"kubernetes.io/projected/1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6-kube-api-access-b6vbq\") pod \"nova-cell0-db-create-bf56t\" (UID: \"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6\") " pod="openstack/nova-cell0-db-create-bf56t" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.790656 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7mcn\" (UniqueName: \"kubernetes.io/projected/9a3dce1e-ade4-4090-a9c5-97238453667f-kube-api-access-k7mcn\") pod \"nova-api-db-create-98rj6\" (UID: \"9a3dce1e-ade4-4090-a9c5-97238453667f\") " pod="openstack/nova-api-db-create-98rj6" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.809196 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7mcn\" (UniqueName: \"kubernetes.io/projected/9a3dce1e-ade4-4090-a9c5-97238453667f-kube-api-access-k7mcn\") pod \"nova-api-db-create-98rj6\" (UID: \"9a3dce1e-ade4-4090-a9c5-97238453667f\") " pod="openstack/nova-api-db-create-98rj6" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.892187 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6vbq\" (UniqueName: \"kubernetes.io/projected/1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6-kube-api-access-b6vbq\") pod \"nova-cell0-db-create-bf56t\" (UID: \"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6\") " pod="openstack/nova-cell0-db-create-bf56t" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.892271 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk4hl\" (UniqueName: \"kubernetes.io/projected/abb77939-b991-416f-be87-085e7d762af6-kube-api-access-zk4hl\") pod \"nova-cell1-db-create-djpqv\" (UID: \"abb77939-b991-416f-be87-085e7d762af6\") " pod="openstack/nova-cell1-db-create-djpqv" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.914670 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6vbq\" (UniqueName: \"kubernetes.io/projected/1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6-kube-api-access-b6vbq\") pod \"nova-cell0-db-create-bf56t\" (UID: \"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6\") " pod="openstack/nova-cell0-db-create-bf56t" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.938877 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98rj6" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.991022 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bf56t" Sep 29 20:41:05 crc kubenswrapper[4741]: I0929 20:41:05.993780 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk4hl\" (UniqueName: \"kubernetes.io/projected/abb77939-b991-416f-be87-085e7d762af6-kube-api-access-zk4hl\") pod \"nova-cell1-db-create-djpqv\" (UID: \"abb77939-b991-416f-be87-085e7d762af6\") " pod="openstack/nova-cell1-db-create-djpqv" Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.010522 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk4hl\" (UniqueName: \"kubernetes.io/projected/abb77939-b991-416f-be87-085e7d762af6-kube-api-access-zk4hl\") pod \"nova-cell1-db-create-djpqv\" (UID: \"abb77939-b991-416f-be87-085e7d762af6\") " pod="openstack/nova-cell1-db-create-djpqv" Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.085656 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:41:06 crc kubenswrapper[4741]: E0929 20:41:06.086192 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.094065 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-djpqv" Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.424727 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-98rj6"] Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.491625 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bf56t"] Sep 29 20:41:06 crc kubenswrapper[4741]: W0929 20:41:06.504069 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f2c8864_ff5d_48f8_bfcd_1f1104ffe9a6.slice/crio-519c0a225f6a67aba95689645151646bf52e918ca7cacaad59a35d784c655fad WatchSource:0}: Error finding container 519c0a225f6a67aba95689645151646bf52e918ca7cacaad59a35d784c655fad: Status 404 returned error can't find the container with id 519c0a225f6a67aba95689645151646bf52e918ca7cacaad59a35d784c655fad Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.555257 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-djpqv"] Sep 29 20:41:06 crc kubenswrapper[4741]: W0929 20:41:06.565102 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabb77939_b991_416f_be87_085e7d762af6.slice/crio-355b90cb9a6d8e1a67cd0dcd416b07e63fda47874e03250459e8f0f8e3265e1f WatchSource:0}: Error finding container 355b90cb9a6d8e1a67cd0dcd416b07e63fda47874e03250459e8f0f8e3265e1f: Status 404 returned error can't find the container with id 355b90cb9a6d8e1a67cd0dcd416b07e63fda47874e03250459e8f0f8e3265e1f Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.663996 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98rj6" event={"ID":"9a3dce1e-ade4-4090-a9c5-97238453667f","Type":"ContainerStarted","Data":"b332dc715b490f8a47542561f3fc98534e0805d24bd11a3cdfed6228dbf21eb6"} Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.664102 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98rj6" event={"ID":"9a3dce1e-ade4-4090-a9c5-97238453667f","Type":"ContainerStarted","Data":"a0bf8365ed5ba4a8e2a940d0308a4fa2e08c63306028ce0e361a868d06677d6e"} Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.666214 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-djpqv" event={"ID":"abb77939-b991-416f-be87-085e7d762af6","Type":"ContainerStarted","Data":"355b90cb9a6d8e1a67cd0dcd416b07e63fda47874e03250459e8f0f8e3265e1f"} Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.668028 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bf56t" event={"ID":"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6","Type":"ContainerStarted","Data":"519c0a225f6a67aba95689645151646bf52e918ca7cacaad59a35d784c655fad"} Sep 29 20:41:06 crc kubenswrapper[4741]: I0929 20:41:06.700169 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-98rj6" podStartSLOduration=1.7001495439999998 podStartE2EDuration="1.700149544s" podCreationTimestamp="2025-09-29 20:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:06.691898859 +0000 UTC m=+5508.339688191" watchObservedRunningTime="2025-09-29 20:41:06.700149544 +0000 UTC m=+5508.347938876" Sep 29 20:41:07 crc kubenswrapper[4741]: I0929 20:41:07.681844 4741 generic.go:334] "Generic (PLEG): container finished" podID="9a3dce1e-ade4-4090-a9c5-97238453667f" containerID="b332dc715b490f8a47542561f3fc98534e0805d24bd11a3cdfed6228dbf21eb6" exitCode=0 Sep 29 20:41:07 crc kubenswrapper[4741]: I0929 20:41:07.681947 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98rj6" event={"ID":"9a3dce1e-ade4-4090-a9c5-97238453667f","Type":"ContainerDied","Data":"b332dc715b490f8a47542561f3fc98534e0805d24bd11a3cdfed6228dbf21eb6"} Sep 29 20:41:07 crc kubenswrapper[4741]: I0929 20:41:07.688063 4741 generic.go:334] "Generic (PLEG): container finished" podID="abb77939-b991-416f-be87-085e7d762af6" containerID="6938ce916b4bebf042498dcf9de1535f9e718aed6efaa9d6936cdae21365486c" exitCode=0 Sep 29 20:41:07 crc kubenswrapper[4741]: I0929 20:41:07.688145 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-djpqv" event={"ID":"abb77939-b991-416f-be87-085e7d762af6","Type":"ContainerDied","Data":"6938ce916b4bebf042498dcf9de1535f9e718aed6efaa9d6936cdae21365486c"} Sep 29 20:41:07 crc kubenswrapper[4741]: I0929 20:41:07.690981 4741 generic.go:334] "Generic (PLEG): container finished" podID="1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6" containerID="2cff54f16bbf7a47c92147596e8cb926ff57cb74f9ca667f207808920f49a0ec" exitCode=0 Sep 29 20:41:07 crc kubenswrapper[4741]: I0929 20:41:07.691013 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bf56t" event={"ID":"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6","Type":"ContainerDied","Data":"2cff54f16bbf7a47c92147596e8cb926ff57cb74f9ca667f207808920f49a0ec"} Sep 29 20:41:07 crc kubenswrapper[4741]: I0929 20:41:07.705255 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-bf56t" podStartSLOduration=2.705239326 podStartE2EDuration="2.705239326s" podCreationTimestamp="2025-09-29 20:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:06.70646206 +0000 UTC m=+5508.354251382" watchObservedRunningTime="2025-09-29 20:41:07.705239326 +0000 UTC m=+5509.353028658" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.123103 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98rj6" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.129326 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-djpqv" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.142108 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bf56t" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.244559 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk4hl\" (UniqueName: \"kubernetes.io/projected/abb77939-b991-416f-be87-085e7d762af6-kube-api-access-zk4hl\") pod \"abb77939-b991-416f-be87-085e7d762af6\" (UID: \"abb77939-b991-416f-be87-085e7d762af6\") " Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.244610 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6vbq\" (UniqueName: \"kubernetes.io/projected/1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6-kube-api-access-b6vbq\") pod \"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6\" (UID: \"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6\") " Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.244642 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7mcn\" (UniqueName: \"kubernetes.io/projected/9a3dce1e-ade4-4090-a9c5-97238453667f-kube-api-access-k7mcn\") pod \"9a3dce1e-ade4-4090-a9c5-97238453667f\" (UID: \"9a3dce1e-ade4-4090-a9c5-97238453667f\") " Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.249340 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3dce1e-ade4-4090-a9c5-97238453667f-kube-api-access-k7mcn" (OuterVolumeSpecName: "kube-api-access-k7mcn") pod "9a3dce1e-ade4-4090-a9c5-97238453667f" (UID: "9a3dce1e-ade4-4090-a9c5-97238453667f"). InnerVolumeSpecName "kube-api-access-k7mcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.249524 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb77939-b991-416f-be87-085e7d762af6-kube-api-access-zk4hl" (OuterVolumeSpecName: "kube-api-access-zk4hl") pod "abb77939-b991-416f-be87-085e7d762af6" (UID: "abb77939-b991-416f-be87-085e7d762af6"). InnerVolumeSpecName "kube-api-access-zk4hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.258120 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6-kube-api-access-b6vbq" (OuterVolumeSpecName: "kube-api-access-b6vbq") pod "1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6" (UID: "1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6"). InnerVolumeSpecName "kube-api-access-b6vbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.348254 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk4hl\" (UniqueName: \"kubernetes.io/projected/abb77939-b991-416f-be87-085e7d762af6-kube-api-access-zk4hl\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.348293 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6vbq\" (UniqueName: \"kubernetes.io/projected/1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6-kube-api-access-b6vbq\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.348307 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7mcn\" (UniqueName: \"kubernetes.io/projected/9a3dce1e-ade4-4090-a9c5-97238453667f-kube-api-access-k7mcn\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.710192 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bf56t" event={"ID":"1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6","Type":"ContainerDied","Data":"519c0a225f6a67aba95689645151646bf52e918ca7cacaad59a35d784c655fad"} Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.710243 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="519c0a225f6a67aba95689645151646bf52e918ca7cacaad59a35d784c655fad" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.710218 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bf56t" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.719651 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98rj6" event={"ID":"9a3dce1e-ade4-4090-a9c5-97238453667f","Type":"ContainerDied","Data":"a0bf8365ed5ba4a8e2a940d0308a4fa2e08c63306028ce0e361a868d06677d6e"} Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.719703 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0bf8365ed5ba4a8e2a940d0308a4fa2e08c63306028ce0e361a868d06677d6e" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.719667 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98rj6" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.723882 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-djpqv" event={"ID":"abb77939-b991-416f-be87-085e7d762af6","Type":"ContainerDied","Data":"355b90cb9a6d8e1a67cd0dcd416b07e63fda47874e03250459e8f0f8e3265e1f"} Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.724242 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-djpqv" Sep 29 20:41:09 crc kubenswrapper[4741]: I0929 20:41:09.724695 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="355b90cb9a6d8e1a67cd0dcd416b07e63fda47874e03250459e8f0f8e3265e1f" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.814704 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-100e-account-create-fd4xg"] Sep 29 20:41:15 crc kubenswrapper[4741]: E0929 20:41:15.815862 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.815889 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: E0929 20:41:15.815909 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3dce1e-ade4-4090-a9c5-97238453667f" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.815921 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3dce1e-ade4-4090-a9c5-97238453667f" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: E0929 20:41:15.815962 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb77939-b991-416f-be87-085e7d762af6" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.815974 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb77939-b991-416f-be87-085e7d762af6" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.816259 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb77939-b991-416f-be87-085e7d762af6" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.816291 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.816313 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3dce1e-ade4-4090-a9c5-97238453667f" containerName="mariadb-database-create" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.817560 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-100e-account-create-fd4xg" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.822884 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.825320 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-100e-account-create-fd4xg"] Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.863348 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxj6c\" (UniqueName: \"kubernetes.io/projected/603b9f56-b1fc-4fdc-bf43-3cc338c737e2-kube-api-access-xxj6c\") pod \"nova-api-100e-account-create-fd4xg\" (UID: \"603b9f56-b1fc-4fdc-bf43-3cc338c737e2\") " pod="openstack/nova-api-100e-account-create-fd4xg" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.965508 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxj6c\" (UniqueName: \"kubernetes.io/projected/603b9f56-b1fc-4fdc-bf43-3cc338c737e2-kube-api-access-xxj6c\") pod \"nova-api-100e-account-create-fd4xg\" (UID: \"603b9f56-b1fc-4fdc-bf43-3cc338c737e2\") " pod="openstack/nova-api-100e-account-create-fd4xg" Sep 29 20:41:15 crc kubenswrapper[4741]: I0929 20:41:15.987430 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxj6c\" (UniqueName: \"kubernetes.io/projected/603b9f56-b1fc-4fdc-bf43-3cc338c737e2-kube-api-access-xxj6c\") pod \"nova-api-100e-account-create-fd4xg\" (UID: \"603b9f56-b1fc-4fdc-bf43-3cc338c737e2\") " pod="openstack/nova-api-100e-account-create-fd4xg" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.020870 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-6963-account-create-f2nw9"] Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.022221 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6963-account-create-f2nw9" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.024020 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.027925 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6963-account-create-f2nw9"] Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.067724 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkjfh\" (UniqueName: \"kubernetes.io/projected/088c9042-fb78-4b36-a9ff-7c5ae4747861-kube-api-access-xkjfh\") pod \"nova-cell0-6963-account-create-f2nw9\" (UID: \"088c9042-fb78-4b36-a9ff-7c5ae4747861\") " pod="openstack/nova-cell0-6963-account-create-f2nw9" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.143066 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-100e-account-create-fd4xg" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.169221 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkjfh\" (UniqueName: \"kubernetes.io/projected/088c9042-fb78-4b36-a9ff-7c5ae4747861-kube-api-access-xkjfh\") pod \"nova-cell0-6963-account-create-f2nw9\" (UID: \"088c9042-fb78-4b36-a9ff-7c5ae4747861\") " pod="openstack/nova-cell0-6963-account-create-f2nw9" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.189447 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkjfh\" (UniqueName: \"kubernetes.io/projected/088c9042-fb78-4b36-a9ff-7c5ae4747861-kube-api-access-xkjfh\") pod \"nova-cell0-6963-account-create-f2nw9\" (UID: \"088c9042-fb78-4b36-a9ff-7c5ae4747861\") " pod="openstack/nova-cell0-6963-account-create-f2nw9" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.211792 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-dd71-account-create-9spsc"] Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.213342 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dd71-account-create-9spsc" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.215332 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.220098 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-dd71-account-create-9spsc"] Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.270731 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnjwd\" (UniqueName: \"kubernetes.io/projected/57de32ae-a8b9-4ab7-b8e5-6cc475543766-kube-api-access-bnjwd\") pod \"nova-cell1-dd71-account-create-9spsc\" (UID: \"57de32ae-a8b9-4ab7-b8e5-6cc475543766\") " pod="openstack/nova-cell1-dd71-account-create-9spsc" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.361533 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6963-account-create-f2nw9" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.371955 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnjwd\" (UniqueName: \"kubernetes.io/projected/57de32ae-a8b9-4ab7-b8e5-6cc475543766-kube-api-access-bnjwd\") pod \"nova-cell1-dd71-account-create-9spsc\" (UID: \"57de32ae-a8b9-4ab7-b8e5-6cc475543766\") " pod="openstack/nova-cell1-dd71-account-create-9spsc" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.390196 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnjwd\" (UniqueName: \"kubernetes.io/projected/57de32ae-a8b9-4ab7-b8e5-6cc475543766-kube-api-access-bnjwd\") pod \"nova-cell1-dd71-account-create-9spsc\" (UID: \"57de32ae-a8b9-4ab7-b8e5-6cc475543766\") " pod="openstack/nova-cell1-dd71-account-create-9spsc" Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.569178 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-100e-account-create-fd4xg"] Sep 29 20:41:16 crc kubenswrapper[4741]: I0929 20:41:16.577734 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dd71-account-create-9spsc" Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:16.801297 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6963-account-create-f2nw9"] Sep 29 20:41:19 crc kubenswrapper[4741]: W0929 20:41:16.804679 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod088c9042_fb78_4b36_a9ff_7c5ae4747861.slice/crio-c9579ca5d87a91e881076657fbf8123c10286db312f3f5916ceb2af7021f77d4 WatchSource:0}: Error finding container c9579ca5d87a91e881076657fbf8123c10286db312f3f5916ceb2af7021f77d4: Status 404 returned error can't find the container with id c9579ca5d87a91e881076657fbf8123c10286db312f3f5916ceb2af7021f77d4 Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:16.805983 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-100e-account-create-fd4xg" event={"ID":"603b9f56-b1fc-4fdc-bf43-3cc338c737e2","Type":"ContainerStarted","Data":"897773d7a29888150348ab3cf089dee033240f61e14a679bba34dd3a0a060a0a"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.026411 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-dd71-account-create-9spsc"] Sep 29 20:41:19 crc kubenswrapper[4741]: W0929 20:41:17.031837 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57de32ae_a8b9_4ab7_b8e5_6cc475543766.slice/crio-07114f1ada33541e9b12b7f6e85fa82fbf2ae377bc327d8ad0246ef552b54e17 WatchSource:0}: Error finding container 07114f1ada33541e9b12b7f6e85fa82fbf2ae377bc327d8ad0246ef552b54e17: Status 404 returned error can't find the container with id 07114f1ada33541e9b12b7f6e85fa82fbf2ae377bc327d8ad0246ef552b54e17 Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.086803 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:41:19 crc kubenswrapper[4741]: E0929 20:41:17.087094 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.817348 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-100e-account-create-fd4xg" event={"ID":"603b9f56-b1fc-4fdc-bf43-3cc338c737e2","Type":"ContainerStarted","Data":"aa0394e84fc4e7abd1d6df3213105e0e1d9f30a587b9fee62406ef5c2ea5761a"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.819330 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dd71-account-create-9spsc" event={"ID":"57de32ae-a8b9-4ab7-b8e5-6cc475543766","Type":"ContainerStarted","Data":"4a662aa200f12dc8733ec49f7e138748b3f8a8dcb88cf0cdbfafa7d7d06b66b8"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.819379 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dd71-account-create-9spsc" event={"ID":"57de32ae-a8b9-4ab7-b8e5-6cc475543766","Type":"ContainerStarted","Data":"07114f1ada33541e9b12b7f6e85fa82fbf2ae377bc327d8ad0246ef552b54e17"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.820824 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6963-account-create-f2nw9" event={"ID":"088c9042-fb78-4b36-a9ff-7c5ae4747861","Type":"ContainerStarted","Data":"4a3567c4a021b49b7eb2de83f7a86b6856c33facca21427f0f7eae32b71679a2"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.820854 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6963-account-create-f2nw9" event={"ID":"088c9042-fb78-4b36-a9ff-7c5ae4747861","Type":"ContainerStarted","Data":"c9579ca5d87a91e881076657fbf8123c10286db312f3f5916ceb2af7021f77d4"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.833843 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-100e-account-create-fd4xg" podStartSLOduration=2.833827259 podStartE2EDuration="2.833827259s" podCreationTimestamp="2025-09-29 20:41:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:17.829447113 +0000 UTC m=+5519.477236445" watchObservedRunningTime="2025-09-29 20:41:17.833827259 +0000 UTC m=+5519.481616591" Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.862009 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-6963-account-create-f2nw9" podStartSLOduration=2.86197886 podStartE2EDuration="2.86197886s" podCreationTimestamp="2025-09-29 20:41:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:17.846714628 +0000 UTC m=+5519.494504050" watchObservedRunningTime="2025-09-29 20:41:17.86197886 +0000 UTC m=+5519.509768232" Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:17.864576 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-dd71-account-create-9spsc" podStartSLOduration=1.864542269 podStartE2EDuration="1.864542269s" podCreationTimestamp="2025-09-29 20:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:17.860452953 +0000 UTC m=+5519.508242285" watchObservedRunningTime="2025-09-29 20:41:17.864542269 +0000 UTC m=+5519.512331611" Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:18.834091 4741 generic.go:334] "Generic (PLEG): container finished" podID="603b9f56-b1fc-4fdc-bf43-3cc338c737e2" containerID="aa0394e84fc4e7abd1d6df3213105e0e1d9f30a587b9fee62406ef5c2ea5761a" exitCode=0 Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:18.834267 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-100e-account-create-fd4xg" event={"ID":"603b9f56-b1fc-4fdc-bf43-3cc338c737e2","Type":"ContainerDied","Data":"aa0394e84fc4e7abd1d6df3213105e0e1d9f30a587b9fee62406ef5c2ea5761a"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:18.839125 4741 generic.go:334] "Generic (PLEG): container finished" podID="57de32ae-a8b9-4ab7-b8e5-6cc475543766" containerID="4a662aa200f12dc8733ec49f7e138748b3f8a8dcb88cf0cdbfafa7d7d06b66b8" exitCode=0 Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:18.839407 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dd71-account-create-9spsc" event={"ID":"57de32ae-a8b9-4ab7-b8e5-6cc475543766","Type":"ContainerDied","Data":"4a662aa200f12dc8733ec49f7e138748b3f8a8dcb88cf0cdbfafa7d7d06b66b8"} Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:18.842524 4741 generic.go:334] "Generic (PLEG): container finished" podID="088c9042-fb78-4b36-a9ff-7c5ae4747861" containerID="4a3567c4a021b49b7eb2de83f7a86b6856c33facca21427f0f7eae32b71679a2" exitCode=0 Sep 29 20:41:19 crc kubenswrapper[4741]: I0929 20:41:18.842559 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6963-account-create-f2nw9" event={"ID":"088c9042-fb78-4b36-a9ff-7c5ae4747861","Type":"ContainerDied","Data":"4a3567c4a021b49b7eb2de83f7a86b6856c33facca21427f0f7eae32b71679a2"} Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.279896 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6963-account-create-f2nw9" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.290462 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-100e-account-create-fd4xg" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.299322 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dd71-account-create-9spsc" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.344281 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxj6c\" (UniqueName: \"kubernetes.io/projected/603b9f56-b1fc-4fdc-bf43-3cc338c737e2-kube-api-access-xxj6c\") pod \"603b9f56-b1fc-4fdc-bf43-3cc338c737e2\" (UID: \"603b9f56-b1fc-4fdc-bf43-3cc338c737e2\") " Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.344360 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkjfh\" (UniqueName: \"kubernetes.io/projected/088c9042-fb78-4b36-a9ff-7c5ae4747861-kube-api-access-xkjfh\") pod \"088c9042-fb78-4b36-a9ff-7c5ae4747861\" (UID: \"088c9042-fb78-4b36-a9ff-7c5ae4747861\") " Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.344407 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnjwd\" (UniqueName: \"kubernetes.io/projected/57de32ae-a8b9-4ab7-b8e5-6cc475543766-kube-api-access-bnjwd\") pod \"57de32ae-a8b9-4ab7-b8e5-6cc475543766\" (UID: \"57de32ae-a8b9-4ab7-b8e5-6cc475543766\") " Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.350149 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57de32ae-a8b9-4ab7-b8e5-6cc475543766-kube-api-access-bnjwd" (OuterVolumeSpecName: "kube-api-access-bnjwd") pod "57de32ae-a8b9-4ab7-b8e5-6cc475543766" (UID: "57de32ae-a8b9-4ab7-b8e5-6cc475543766"). InnerVolumeSpecName "kube-api-access-bnjwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.350636 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088c9042-fb78-4b36-a9ff-7c5ae4747861-kube-api-access-xkjfh" (OuterVolumeSpecName: "kube-api-access-xkjfh") pod "088c9042-fb78-4b36-a9ff-7c5ae4747861" (UID: "088c9042-fb78-4b36-a9ff-7c5ae4747861"). InnerVolumeSpecName "kube-api-access-xkjfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.350782 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603b9f56-b1fc-4fdc-bf43-3cc338c737e2-kube-api-access-xxj6c" (OuterVolumeSpecName: "kube-api-access-xxj6c") pod "603b9f56-b1fc-4fdc-bf43-3cc338c737e2" (UID: "603b9f56-b1fc-4fdc-bf43-3cc338c737e2"). InnerVolumeSpecName "kube-api-access-xxj6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.445787 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkjfh\" (UniqueName: \"kubernetes.io/projected/088c9042-fb78-4b36-a9ff-7c5ae4747861-kube-api-access-xkjfh\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.445827 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnjwd\" (UniqueName: \"kubernetes.io/projected/57de32ae-a8b9-4ab7-b8e5-6cc475543766-kube-api-access-bnjwd\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.445836 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxj6c\" (UniqueName: \"kubernetes.io/projected/603b9f56-b1fc-4fdc-bf43-3cc338c737e2-kube-api-access-xxj6c\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.888177 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-100e-account-create-fd4xg" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.888203 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-100e-account-create-fd4xg" event={"ID":"603b9f56-b1fc-4fdc-bf43-3cc338c737e2","Type":"ContainerDied","Data":"897773d7a29888150348ab3cf089dee033240f61e14a679bba34dd3a0a060a0a"} Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.888697 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="897773d7a29888150348ab3cf089dee033240f61e14a679bba34dd3a0a060a0a" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.890368 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dd71-account-create-9spsc" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.890430 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dd71-account-create-9spsc" event={"ID":"57de32ae-a8b9-4ab7-b8e5-6cc475543766","Type":"ContainerDied","Data":"07114f1ada33541e9b12b7f6e85fa82fbf2ae377bc327d8ad0246ef552b54e17"} Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.890479 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07114f1ada33541e9b12b7f6e85fa82fbf2ae377bc327d8ad0246ef552b54e17" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.892041 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6963-account-create-f2nw9" event={"ID":"088c9042-fb78-4b36-a9ff-7c5ae4747861","Type":"ContainerDied","Data":"c9579ca5d87a91e881076657fbf8123c10286db312f3f5916ceb2af7021f77d4"} Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.892069 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9579ca5d87a91e881076657fbf8123c10286db312f3f5916ceb2af7021f77d4" Sep 29 20:41:20 crc kubenswrapper[4741]: I0929 20:41:20.892166 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6963-account-create-f2nw9" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.260381 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzs57"] Sep 29 20:41:21 crc kubenswrapper[4741]: E0929 20:41:21.260841 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57de32ae-a8b9-4ab7-b8e5-6cc475543766" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.260865 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="57de32ae-a8b9-4ab7-b8e5-6cc475543766" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: E0929 20:41:21.260884 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603b9f56-b1fc-4fdc-bf43-3cc338c737e2" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.260919 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="603b9f56-b1fc-4fdc-bf43-3cc338c737e2" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: E0929 20:41:21.260951 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088c9042-fb78-4b36-a9ff-7c5ae4747861" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.260960 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="088c9042-fb78-4b36-a9ff-7c5ae4747861" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.261159 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="603b9f56-b1fc-4fdc-bf43-3cc338c737e2" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.261185 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="088c9042-fb78-4b36-a9ff-7c5ae4747861" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.261197 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="57de32ae-a8b9-4ab7-b8e5-6cc475543766" containerName="mariadb-account-create" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.261935 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.263888 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.263959 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.264401 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-phz8q" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.272996 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzs57"] Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.363619 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jprlv\" (UniqueName: \"kubernetes.io/projected/089efa06-d16d-4604-8d3e-6f7b2c427dd8-kube-api-access-jprlv\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.363712 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-scripts\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.363852 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.363923 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-config-data\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.466591 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-scripts\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.466723 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.466773 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-config-data\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.466808 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jprlv\" (UniqueName: \"kubernetes.io/projected/089efa06-d16d-4604-8d3e-6f7b2c427dd8-kube-api-access-jprlv\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.473073 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-config-data\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.475531 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-scripts\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.478302 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.488302 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jprlv\" (UniqueName: \"kubernetes.io/projected/089efa06-d16d-4604-8d3e-6f7b2c427dd8-kube-api-access-jprlv\") pod \"nova-cell0-conductor-db-sync-jzs57\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:21 crc kubenswrapper[4741]: I0929 20:41:21.596134 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:22 crc kubenswrapper[4741]: I0929 20:41:22.048433 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzs57"] Sep 29 20:41:22 crc kubenswrapper[4741]: I0929 20:41:22.912295 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzs57" event={"ID":"089efa06-d16d-4604-8d3e-6f7b2c427dd8","Type":"ContainerStarted","Data":"486fb5c1b7e1d8ae7fdf980d6a6ce27ee1ee415f2e3560964ab4e4ad889197a2"} Sep 29 20:41:22 crc kubenswrapper[4741]: I0929 20:41:22.914085 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzs57" event={"ID":"089efa06-d16d-4604-8d3e-6f7b2c427dd8","Type":"ContainerStarted","Data":"5231c40da23977c11d3432724c9eb3a5dca78fab57bc96ab054522556901a5c4"} Sep 29 20:41:22 crc kubenswrapper[4741]: I0929 20:41:22.935110 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-jzs57" podStartSLOduration=1.935093815 podStartE2EDuration="1.935093815s" podCreationTimestamp="2025-09-29 20:41:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:22.930493213 +0000 UTC m=+5524.578282615" watchObservedRunningTime="2025-09-29 20:41:22.935093815 +0000 UTC m=+5524.582883147" Sep 29 20:41:27 crc kubenswrapper[4741]: I0929 20:41:27.151121 4741 scope.go:117] "RemoveContainer" containerID="87771f9128e30a94e6a5ed3e5a856844569a501687ca316d95f545805fa3b759" Sep 29 20:41:28 crc kubenswrapper[4741]: I0929 20:41:28.085659 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:41:28 crc kubenswrapper[4741]: E0929 20:41:28.085887 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:41:33 crc kubenswrapper[4741]: I0929 20:41:33.023515 4741 generic.go:334] "Generic (PLEG): container finished" podID="089efa06-d16d-4604-8d3e-6f7b2c427dd8" containerID="486fb5c1b7e1d8ae7fdf980d6a6ce27ee1ee415f2e3560964ab4e4ad889197a2" exitCode=0 Sep 29 20:41:33 crc kubenswrapper[4741]: I0929 20:41:33.023631 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzs57" event={"ID":"089efa06-d16d-4604-8d3e-6f7b2c427dd8","Type":"ContainerDied","Data":"486fb5c1b7e1d8ae7fdf980d6a6ce27ee1ee415f2e3560964ab4e4ad889197a2"} Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.373901 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.505386 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jprlv\" (UniqueName: \"kubernetes.io/projected/089efa06-d16d-4604-8d3e-6f7b2c427dd8-kube-api-access-jprlv\") pod \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.505567 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-config-data\") pod \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.506433 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-combined-ca-bundle\") pod \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.506519 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-scripts\") pod \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\" (UID: \"089efa06-d16d-4604-8d3e-6f7b2c427dd8\") " Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.511581 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089efa06-d16d-4604-8d3e-6f7b2c427dd8-kube-api-access-jprlv" (OuterVolumeSpecName: "kube-api-access-jprlv") pod "089efa06-d16d-4604-8d3e-6f7b2c427dd8" (UID: "089efa06-d16d-4604-8d3e-6f7b2c427dd8"). InnerVolumeSpecName "kube-api-access-jprlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.511652 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-scripts" (OuterVolumeSpecName: "scripts") pod "089efa06-d16d-4604-8d3e-6f7b2c427dd8" (UID: "089efa06-d16d-4604-8d3e-6f7b2c427dd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.530536 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-config-data" (OuterVolumeSpecName: "config-data") pod "089efa06-d16d-4604-8d3e-6f7b2c427dd8" (UID: "089efa06-d16d-4604-8d3e-6f7b2c427dd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.531677 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "089efa06-d16d-4604-8d3e-6f7b2c427dd8" (UID: "089efa06-d16d-4604-8d3e-6f7b2c427dd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.608606 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jprlv\" (UniqueName: \"kubernetes.io/projected/089efa06-d16d-4604-8d3e-6f7b2c427dd8-kube-api-access-jprlv\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.608641 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.608650 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:34 crc kubenswrapper[4741]: I0929 20:41:34.608657 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089efa06-d16d-4604-8d3e-6f7b2c427dd8-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.043686 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzs57" event={"ID":"089efa06-d16d-4604-8d3e-6f7b2c427dd8","Type":"ContainerDied","Data":"5231c40da23977c11d3432724c9eb3a5dca78fab57bc96ab054522556901a5c4"} Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.043762 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5231c40da23977c11d3432724c9eb3a5dca78fab57bc96ab054522556901a5c4" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.043757 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzs57" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.125942 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:41:35 crc kubenswrapper[4741]: E0929 20:41:35.126434 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089efa06-d16d-4604-8d3e-6f7b2c427dd8" containerName="nova-cell0-conductor-db-sync" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.126458 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="089efa06-d16d-4604-8d3e-6f7b2c427dd8" containerName="nova-cell0-conductor-db-sync" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.126882 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="089efa06-d16d-4604-8d3e-6f7b2c427dd8" containerName="nova-cell0-conductor-db-sync" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.127825 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.132959 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-phz8q" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.133138 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.144721 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.321064 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6dtq\" (UniqueName: \"kubernetes.io/projected/11c6a275-bb37-4f95-a582-2406799d11c1-kube-api-access-x6dtq\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.321142 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.321234 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.423319 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.423444 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6dtq\" (UniqueName: \"kubernetes.io/projected/11c6a275-bb37-4f95-a582-2406799d11c1-kube-api-access-x6dtq\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.423512 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.427753 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.427851 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.441669 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6dtq\" (UniqueName: \"kubernetes.io/projected/11c6a275-bb37-4f95-a582-2406799d11c1-kube-api-access-x6dtq\") pod \"nova-cell0-conductor-0\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.456007 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:35 crc kubenswrapper[4741]: I0929 20:41:35.915516 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:41:35 crc kubenswrapper[4741]: W0929 20:41:35.920703 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11c6a275_bb37_4f95_a582_2406799d11c1.slice/crio-0deb5ab5ecfcae38c9a5fa92818d47187501ba22efacf716d388f200bae48fb6 WatchSource:0}: Error finding container 0deb5ab5ecfcae38c9a5fa92818d47187501ba22efacf716d388f200bae48fb6: Status 404 returned error can't find the container with id 0deb5ab5ecfcae38c9a5fa92818d47187501ba22efacf716d388f200bae48fb6 Sep 29 20:41:36 crc kubenswrapper[4741]: I0929 20:41:36.069165 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"11c6a275-bb37-4f95-a582-2406799d11c1","Type":"ContainerStarted","Data":"0deb5ab5ecfcae38c9a5fa92818d47187501ba22efacf716d388f200bae48fb6"} Sep 29 20:41:37 crc kubenswrapper[4741]: I0929 20:41:37.078601 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"11c6a275-bb37-4f95-a582-2406799d11c1","Type":"ContainerStarted","Data":"bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c"} Sep 29 20:41:37 crc kubenswrapper[4741]: I0929 20:41:37.078727 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:37 crc kubenswrapper[4741]: I0929 20:41:37.102881 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.102861777 podStartE2EDuration="2.102861777s" podCreationTimestamp="2025-09-29 20:41:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:37.09550657 +0000 UTC m=+5538.743295902" watchObservedRunningTime="2025-09-29 20:41:37.102861777 +0000 UTC m=+5538.750651109" Sep 29 20:41:39 crc kubenswrapper[4741]: I0929 20:41:39.093339 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:41:40 crc kubenswrapper[4741]: I0929 20:41:40.110786 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"564907e3fd83739b2ba2e94f48864f64d35270919df4a735eaebeff724bee01a"} Sep 29 20:41:45 crc kubenswrapper[4741]: I0929 20:41:45.497187 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.076649 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-84tzh"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.078031 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.079789 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.079946 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.102070 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-84tzh"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.206459 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.208156 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.212788 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.213490 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v4tk\" (UniqueName: \"kubernetes.io/projected/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-kube-api-access-6v4tk\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.213793 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-scripts\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.213864 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-config-data\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.213921 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.221705 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.268009 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.272423 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.275963 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.286514 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.311498 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.312607 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.317744 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319360 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7gv2\" (UniqueName: \"kubernetes.io/projected/a3aafbc7-5b56-4632-97ab-002cc11f4e19-kube-api-access-j7gv2\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319437 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-config-data\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319468 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319499 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-config-data\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319518 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319576 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v4tk\" (UniqueName: \"kubernetes.io/projected/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-kube-api-access-6v4tk\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319604 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3aafbc7-5b56-4632-97ab-002cc11f4e19-logs\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319622 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319641 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319660 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-scripts\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.319674 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmjfg\" (UniqueName: \"kubernetes.io/projected/cb20babd-3a68-4c55-a6f6-2da696e4044a-kube-api-access-fmjfg\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.325712 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-config-data\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.327516 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.334988 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.340932 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-scripts\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.346010 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v4tk\" (UniqueName: \"kubernetes.io/projected/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-kube-api-access-6v4tk\") pod \"nova-cell0-cell-mapping-84tzh\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.373165 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.374610 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.379788 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.383567 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.399719 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421551 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421616 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421661 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-config-data\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421682 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-logs\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421721 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-config-data\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421783 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3aafbc7-5b56-4632-97ab-002cc11f4e19-logs\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421830 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421848 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lghc8\" (UniqueName: \"kubernetes.io/projected/cbc6695d-ea89-43fe-80e4-b19237977890-kube-api-access-lghc8\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421868 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421886 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-config-data\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421909 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421932 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmjfg\" (UniqueName: \"kubernetes.io/projected/cb20babd-3a68-4c55-a6f6-2da696e4044a-kube-api-access-fmjfg\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421959 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzbm6\" (UniqueName: \"kubernetes.io/projected/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-kube-api-access-xzbm6\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.421988 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7gv2\" (UniqueName: \"kubernetes.io/projected/a3aafbc7-5b56-4632-97ab-002cc11f4e19-kube-api-access-j7gv2\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.423938 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3aafbc7-5b56-4632-97ab-002cc11f4e19-logs\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.427711 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-config-data\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.428444 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.436662 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.440496 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.445510 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7gv2\" (UniqueName: \"kubernetes.io/projected/a3aafbc7-5b56-4632-97ab-002cc11f4e19-kube-api-access-j7gv2\") pod \"nova-api-0\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.455425 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmjfg\" (UniqueName: \"kubernetes.io/projected/cb20babd-3a68-4c55-a6f6-2da696e4044a-kube-api-access-fmjfg\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.458209 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d56b878c9-4pvdq"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.462496 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.483498 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d56b878c9-4pvdq"] Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523441 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-dns-svc\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523514 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lghc8\" (UniqueName: \"kubernetes.io/projected/cbc6695d-ea89-43fe-80e4-b19237977890-kube-api-access-lghc8\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523542 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523563 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-config-data\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523598 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgf4s\" (UniqueName: \"kubernetes.io/projected/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-kube-api-access-qgf4s\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523637 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzbm6\" (UniqueName: \"kubernetes.io/projected/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-kube-api-access-xzbm6\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523668 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-sb\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523700 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-config\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523738 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-nb\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523781 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523847 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-logs\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.523895 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-config-data\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.527765 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-logs\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.530078 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.537125 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-config-data\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.537205 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.537735 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-config-data\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.538260 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.544032 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lghc8\" (UniqueName: \"kubernetes.io/projected/cbc6695d-ea89-43fe-80e4-b19237977890-kube-api-access-lghc8\") pod \"nova-scheduler-0\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.551065 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzbm6\" (UniqueName: \"kubernetes.io/projected/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-kube-api-access-xzbm6\") pod \"nova-metadata-0\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.597265 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.625425 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgf4s\" (UniqueName: \"kubernetes.io/projected/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-kube-api-access-qgf4s\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.625498 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-sb\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.625526 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-config\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.625553 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-nb\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.625654 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-dns-svc\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.628703 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-nb\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.629220 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-config\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.629729 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-dns-svc\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.647855 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-sb\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.663674 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgf4s\" (UniqueName: \"kubernetes.io/projected/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-kube-api-access-qgf4s\") pod \"dnsmasq-dns-7d56b878c9-4pvdq\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.728939 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.843298 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.903699 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:46 crc kubenswrapper[4741]: I0929 20:41:46.994520 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-84tzh"] Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.136350 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:47 crc kubenswrapper[4741]: W0929 20:41:47.153155 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3aafbc7_5b56_4632_97ab_002cc11f4e19.slice/crio-674bc4ea7fd8bfb7f587e6926750c766236218d3da6bfb550e68d43336aa78ea WatchSource:0}: Error finding container 674bc4ea7fd8bfb7f587e6926750c766236218d3da6bfb550e68d43336aa78ea: Status 404 returned error can't find the container with id 674bc4ea7fd8bfb7f587e6926750c766236218d3da6bfb550e68d43336aa78ea Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.220222 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.221799 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3aafbc7-5b56-4632-97ab-002cc11f4e19","Type":"ContainerStarted","Data":"674bc4ea7fd8bfb7f587e6926750c766236218d3da6bfb550e68d43336aa78ea"} Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.223230 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-84tzh" event={"ID":"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e","Type":"ContainerStarted","Data":"f1d12b602433daa91a1508dde662aeb7ed93a482d3424c2527eb7e1cecf2e201"} Sep 29 20:41:47 crc kubenswrapper[4741]: W0929 20:41:47.223676 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb20babd_3a68_4c55_a6f6_2da696e4044a.slice/crio-8a39538de64cda6d6be93145893702e399ae3193ab518c9c65cb3d90597ccc79 WatchSource:0}: Error finding container 8a39538de64cda6d6be93145893702e399ae3193ab518c9c65cb3d90597ccc79: Status 404 returned error can't find the container with id 8a39538de64cda6d6be93145893702e399ae3193ab518c9c65cb3d90597ccc79 Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.318428 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:47 crc kubenswrapper[4741]: W0929 20:41:47.319902 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbc6695d_ea89_43fe_80e4_b19237977890.slice/crio-b1c0d232b57dee11e15874d09854c48c213d6e9092593cc60c7525e0a8e7f8f1 WatchSource:0}: Error finding container b1c0d232b57dee11e15874d09854c48c213d6e9092593cc60c7525e0a8e7f8f1: Status 404 returned error can't find the container with id b1c0d232b57dee11e15874d09854c48c213d6e9092593cc60c7525e0a8e7f8f1 Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.411231 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.482893 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d56b878c9-4pvdq"] Sep 29 20:41:47 crc kubenswrapper[4741]: W0929 20:41:47.490521 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeab69ab2_6d74_4862_9a59_42fa5c09ccaa.slice/crio-0f4092149aad532f4c8df961b8e8082c8def6078230a2ea1be021430931588ea WatchSource:0}: Error finding container 0f4092149aad532f4c8df961b8e8082c8def6078230a2ea1be021430931588ea: Status 404 returned error can't find the container with id 0f4092149aad532f4c8df961b8e8082c8def6078230a2ea1be021430931588ea Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.563609 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd5mn"] Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.577217 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.584186 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd5mn"] Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.585932 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.587864 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.646074 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.646170 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-config-data\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.646436 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxjsw\" (UniqueName: \"kubernetes.io/projected/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-kube-api-access-rxjsw\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.646629 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-scripts\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.748853 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxjsw\" (UniqueName: \"kubernetes.io/projected/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-kube-api-access-rxjsw\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.749027 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-scripts\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.749063 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.749088 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-config-data\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.765848 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-scripts\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.765935 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.767067 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-config-data\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.770736 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxjsw\" (UniqueName: \"kubernetes.io/projected/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-kube-api-access-rxjsw\") pod \"nova-cell1-conductor-db-sync-rd5mn\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:47 crc kubenswrapper[4741]: I0929 20:41:47.995940 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.244169 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad","Type":"ContainerStarted","Data":"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.244840 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad","Type":"ContainerStarted","Data":"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.244858 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad","Type":"ContainerStarted","Data":"0505f1bb4193fea0f39344b1b102a26551f1b7a708838d2a52d90f1353f0996e"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.259563 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3aafbc7-5b56-4632-97ab-002cc11f4e19","Type":"ContainerStarted","Data":"c309dadd82316dcb4ff62b6317fcdba8e66506f29bcd43681d01c70241afbbbc"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.259615 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3aafbc7-5b56-4632-97ab-002cc11f4e19","Type":"ContainerStarted","Data":"7f0f1b826d8acc9a4744a812ca4ae4a1606e3a4b1503ef22b8ae5795805a8ee9"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.276215 4741 generic.go:334] "Generic (PLEG): container finished" podID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerID="119d985cb03d25915f1ec0f93863619c09e7b456672ca6d486d3fefdac475d82" exitCode=0 Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.276319 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" event={"ID":"eab69ab2-6d74-4862-9a59-42fa5c09ccaa","Type":"ContainerDied","Data":"119d985cb03d25915f1ec0f93863619c09e7b456672ca6d486d3fefdac475d82"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.276374 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" event={"ID":"eab69ab2-6d74-4862-9a59-42fa5c09ccaa","Type":"ContainerStarted","Data":"0f4092149aad532f4c8df961b8e8082c8def6078230a2ea1be021430931588ea"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.282899 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-84tzh" event={"ID":"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e","Type":"ContainerStarted","Data":"6ad2d5b3090e1a80a515918e076f42c30f5497305ee46097585bb4506ae16479"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.291596 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.291571827 podStartE2EDuration="2.291571827s" podCreationTimestamp="2025-09-29 20:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:48.267825421 +0000 UTC m=+5549.915614753" watchObservedRunningTime="2025-09-29 20:41:48.291571827 +0000 UTC m=+5549.939361159" Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.300335 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.300313518 podStartE2EDuration="2.300313518s" podCreationTimestamp="2025-09-29 20:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:48.285521109 +0000 UTC m=+5549.933310451" watchObservedRunningTime="2025-09-29 20:41:48.300313518 +0000 UTC m=+5549.948102860" Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.313948 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbc6695d-ea89-43fe-80e4-b19237977890","Type":"ContainerStarted","Data":"e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.314001 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbc6695d-ea89-43fe-80e4-b19237977890","Type":"ContainerStarted","Data":"b1c0d232b57dee11e15874d09854c48c213d6e9092593cc60c7525e0a8e7f8f1"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.341960 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb20babd-3a68-4c55-a6f6-2da696e4044a","Type":"ContainerStarted","Data":"6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.342004 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb20babd-3a68-4c55-a6f6-2da696e4044a","Type":"ContainerStarted","Data":"8a39538de64cda6d6be93145893702e399ae3193ab518c9c65cb3d90597ccc79"} Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.351787 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-84tzh" podStartSLOduration=2.3517633 podStartE2EDuration="2.3517633s" podCreationTimestamp="2025-09-29 20:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:48.322847285 +0000 UTC m=+5549.970636617" watchObservedRunningTime="2025-09-29 20:41:48.3517633 +0000 UTC m=+5549.999552632" Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.372540 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.372518912 podStartE2EDuration="2.372518912s" podCreationTimestamp="2025-09-29 20:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:48.347611512 +0000 UTC m=+5549.995400844" watchObservedRunningTime="2025-09-29 20:41:48.372518912 +0000 UTC m=+5550.020308244" Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.386959 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.386936989 podStartE2EDuration="2.386936989s" podCreationTimestamp="2025-09-29 20:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:48.360645935 +0000 UTC m=+5550.008435267" watchObservedRunningTime="2025-09-29 20:41:48.386936989 +0000 UTC m=+5550.034726321" Sep 29 20:41:48 crc kubenswrapper[4741]: I0929 20:41:48.510347 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd5mn"] Sep 29 20:41:48 crc kubenswrapper[4741]: W0929 20:41:48.522038 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89fed09c_7c9c_40c8_82e3_dca8f90f0ff2.slice/crio-9263fd504c23ea2b509a5d3ba65587d021d034a1fb65ec42edd7d9a83a1d899c WatchSource:0}: Error finding container 9263fd504c23ea2b509a5d3ba65587d021d034a1fb65ec42edd7d9a83a1d899c: Status 404 returned error can't find the container with id 9263fd504c23ea2b509a5d3ba65587d021d034a1fb65ec42edd7d9a83a1d899c Sep 29 20:41:49 crc kubenswrapper[4741]: I0929 20:41:49.359938 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" event={"ID":"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2","Type":"ContainerStarted","Data":"9e372166fb4936206ea260fd5aeea6e1ba893242dd5e5122802e936310050c73"} Sep 29 20:41:49 crc kubenswrapper[4741]: I0929 20:41:49.360321 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" event={"ID":"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2","Type":"ContainerStarted","Data":"9263fd504c23ea2b509a5d3ba65587d021d034a1fb65ec42edd7d9a83a1d899c"} Sep 29 20:41:49 crc kubenswrapper[4741]: I0929 20:41:49.372475 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" event={"ID":"eab69ab2-6d74-4862-9a59-42fa5c09ccaa","Type":"ContainerStarted","Data":"91d3646c187680b47e8a22e392a96da2a25cdea71c527b0f773fed6761b5a71a"} Sep 29 20:41:49 crc kubenswrapper[4741]: I0929 20:41:49.388595 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" podStartSLOduration=2.388568554 podStartE2EDuration="2.388568554s" podCreationTimestamp="2025-09-29 20:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:49.386308884 +0000 UTC m=+5551.034098226" watchObservedRunningTime="2025-09-29 20:41:49.388568554 +0000 UTC m=+5551.036357886" Sep 29 20:41:49 crc kubenswrapper[4741]: I0929 20:41:49.415747 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" podStartSLOduration=3.415721224 podStartE2EDuration="3.415721224s" podCreationTimestamp="2025-09-29 20:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:49.412895296 +0000 UTC m=+5551.060684668" watchObservedRunningTime="2025-09-29 20:41:49.415721224 +0000 UTC m=+5551.063510586" Sep 29 20:41:50 crc kubenswrapper[4741]: I0929 20:41:50.381439 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:51 crc kubenswrapper[4741]: I0929 20:41:51.598382 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:51 crc kubenswrapper[4741]: I0929 20:41:51.730286 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 20:41:51 crc kubenswrapper[4741]: I0929 20:41:51.844076 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:41:51 crc kubenswrapper[4741]: I0929 20:41:51.844147 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:41:52 crc kubenswrapper[4741]: I0929 20:41:52.414059 4741 generic.go:334] "Generic (PLEG): container finished" podID="bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" containerID="6ad2d5b3090e1a80a515918e076f42c30f5497305ee46097585bb4506ae16479" exitCode=0 Sep 29 20:41:52 crc kubenswrapper[4741]: I0929 20:41:52.414123 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-84tzh" event={"ID":"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e","Type":"ContainerDied","Data":"6ad2d5b3090e1a80a515918e076f42c30f5497305ee46097585bb4506ae16479"} Sep 29 20:41:53 crc kubenswrapper[4741]: I0929 20:41:53.837277 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:53 crc kubenswrapper[4741]: I0929 20:41:53.981301 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-combined-ca-bundle\") pod \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " Sep 29 20:41:53 crc kubenswrapper[4741]: I0929 20:41:53.981732 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-scripts\") pod \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " Sep 29 20:41:53 crc kubenswrapper[4741]: I0929 20:41:53.981890 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-config-data\") pod \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " Sep 29 20:41:53 crc kubenswrapper[4741]: I0929 20:41:53.981987 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v4tk\" (UniqueName: \"kubernetes.io/projected/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-kube-api-access-6v4tk\") pod \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\" (UID: \"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e\") " Sep 29 20:41:53 crc kubenswrapper[4741]: I0929 20:41:53.997549 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-scripts" (OuterVolumeSpecName: "scripts") pod "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" (UID: "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:53 crc kubenswrapper[4741]: I0929 20:41:53.998474 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-kube-api-access-6v4tk" (OuterVolumeSpecName: "kube-api-access-6v4tk") pod "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" (UID: "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e"). InnerVolumeSpecName "kube-api-access-6v4tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.007609 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-config-data" (OuterVolumeSpecName: "config-data") pod "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" (UID: "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.008960 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" (UID: "bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.084146 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.084249 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.084299 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.084347 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v4tk\" (UniqueName: \"kubernetes.io/projected/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e-kube-api-access-6v4tk\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.438720 4741 generic.go:334] "Generic (PLEG): container finished" podID="89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" containerID="9e372166fb4936206ea260fd5aeea6e1ba893242dd5e5122802e936310050c73" exitCode=0 Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.438816 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" event={"ID":"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2","Type":"ContainerDied","Data":"9e372166fb4936206ea260fd5aeea6e1ba893242dd5e5122802e936310050c73"} Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.441188 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-84tzh" event={"ID":"bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e","Type":"ContainerDied","Data":"f1d12b602433daa91a1508dde662aeb7ed93a482d3424c2527eb7e1cecf2e201"} Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.441220 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1d12b602433daa91a1508dde662aeb7ed93a482d3424c2527eb7e1cecf2e201" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.441645 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-84tzh" Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.628551 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.628975 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-log" containerID="cri-o://7f0f1b826d8acc9a4744a812ca4ae4a1606e3a4b1503ef22b8ae5795805a8ee9" gracePeriod=30 Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.629246 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-api" containerID="cri-o://c309dadd82316dcb4ff62b6317fcdba8e66506f29bcd43681d01c70241afbbbc" gracePeriod=30 Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.718903 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.719116 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cbc6695d-ea89-43fe-80e4-b19237977890" containerName="nova-scheduler-scheduler" containerID="cri-o://e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11" gracePeriod=30 Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.734766 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.735234 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-log" containerID="cri-o://b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3" gracePeriod=30 Sep 29 20:41:54 crc kubenswrapper[4741]: I0929 20:41:54.735343 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-metadata" containerID="cri-o://c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f" gracePeriod=30 Sep 29 20:41:55 crc kubenswrapper[4741]: E0929 20:41:55.210464 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3aafbc7_5b56_4632_97ab_002cc11f4e19.slice/crio-conmon-c309dadd82316dcb4ff62b6317fcdba8e66506f29bcd43681d01c70241afbbbc.scope\": RecentStats: unable to find data in memory cache]" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.364535 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.498545 4741 generic.go:334] "Generic (PLEG): container finished" podID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerID="c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f" exitCode=0 Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.498581 4741 generic.go:334] "Generic (PLEG): container finished" podID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerID="b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3" exitCode=143 Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.498738 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.499456 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad","Type":"ContainerDied","Data":"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f"} Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.499492 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad","Type":"ContainerDied","Data":"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3"} Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.499505 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad","Type":"ContainerDied","Data":"0505f1bb4193fea0f39344b1b102a26551f1b7a708838d2a52d90f1353f0996e"} Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.499519 4741 scope.go:117] "RemoveContainer" containerID="c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.517570 4741 generic.go:334] "Generic (PLEG): container finished" podID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerID="c309dadd82316dcb4ff62b6317fcdba8e66506f29bcd43681d01c70241afbbbc" exitCode=0 Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.517602 4741 generic.go:334] "Generic (PLEG): container finished" podID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerID="7f0f1b826d8acc9a4744a812ca4ae4a1606e3a4b1503ef22b8ae5795805a8ee9" exitCode=143 Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.517808 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3aafbc7-5b56-4632-97ab-002cc11f4e19","Type":"ContainerDied","Data":"c309dadd82316dcb4ff62b6317fcdba8e66506f29bcd43681d01c70241afbbbc"} Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.517834 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3aafbc7-5b56-4632-97ab-002cc11f4e19","Type":"ContainerDied","Data":"7f0f1b826d8acc9a4744a812ca4ae4a1606e3a4b1503ef22b8ae5795805a8ee9"} Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.530945 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-combined-ca-bundle\") pod \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.531150 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-logs\") pod \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.531186 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-config-data\") pod \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.531712 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzbm6\" (UniqueName: \"kubernetes.io/projected/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-kube-api-access-xzbm6\") pod \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\" (UID: \"776b8de5-037e-4cc6-9cb4-d7cefa25f7ad\") " Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.531644 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-logs" (OuterVolumeSpecName: "logs") pod "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" (UID: "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.532432 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.551710 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-kube-api-access-xzbm6" (OuterVolumeSpecName: "kube-api-access-xzbm6") pod "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" (UID: "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad"). InnerVolumeSpecName "kube-api-access-xzbm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.568957 4741 scope.go:117] "RemoveContainer" containerID="b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.594770 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-config-data" (OuterVolumeSpecName: "config-data") pod "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" (UID: "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.607510 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" (UID: "776b8de5-037e-4cc6-9cb4-d7cefa25f7ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.636247 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzbm6\" (UniqueName: \"kubernetes.io/projected/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-kube-api-access-xzbm6\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.636285 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.636298 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.674454 4741 scope.go:117] "RemoveContainer" containerID="c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f" Sep 29 20:41:55 crc kubenswrapper[4741]: E0929 20:41:55.678498 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f\": container with ID starting with c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f not found: ID does not exist" containerID="c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.678562 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f"} err="failed to get container status \"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f\": rpc error: code = NotFound desc = could not find container \"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f\": container with ID starting with c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f not found: ID does not exist" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.678591 4741 scope.go:117] "RemoveContainer" containerID="b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3" Sep 29 20:41:55 crc kubenswrapper[4741]: E0929 20:41:55.678903 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3\": container with ID starting with b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3 not found: ID does not exist" containerID="b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.678947 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3"} err="failed to get container status \"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3\": rpc error: code = NotFound desc = could not find container \"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3\": container with ID starting with b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3 not found: ID does not exist" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.678964 4741 scope.go:117] "RemoveContainer" containerID="c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.679438 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f"} err="failed to get container status \"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f\": rpc error: code = NotFound desc = could not find container \"c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f\": container with ID starting with c32c5f0b34314570c0b692334a524a0a4dd63f1aa4141098ad6062b46e6d9d3f not found: ID does not exist" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.679461 4741 scope.go:117] "RemoveContainer" containerID="b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.679999 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3"} err="failed to get container status \"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3\": rpc error: code = NotFound desc = could not find container \"b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3\": container with ID starting with b2337fc07e0f5ed369fb507865f5ee0e5f45680dcc0279bd4cadb8bba54f3fb3 not found: ID does not exist" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.849536 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.856944 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.899078 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:55 crc kubenswrapper[4741]: E0929 20:41:55.899578 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-metadata" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.899597 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-metadata" Sep 29 20:41:55 crc kubenswrapper[4741]: E0929 20:41:55.899620 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-log" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.899628 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-log" Sep 29 20:41:55 crc kubenswrapper[4741]: E0929 20:41:55.899660 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" containerName="nova-manage" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.899668 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" containerName="nova-manage" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.899875 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-metadata" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.899893 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" containerName="nova-metadata-log" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.899904 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" containerName="nova-manage" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.901041 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.903935 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 20:41:55 crc kubenswrapper[4741]: I0929 20:41:55.912697 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.043845 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-logs\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.043932 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj5v2\" (UniqueName: \"kubernetes.io/projected/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-kube-api-access-xj5v2\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.044011 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-config-data\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.044041 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.158107 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-logs\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.158509 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj5v2\" (UniqueName: \"kubernetes.io/projected/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-kube-api-access-xj5v2\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.158574 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-config-data\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.158612 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.159873 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-logs\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.164215 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.165849 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-config-data\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.173422 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.175636 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.178550 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj5v2\" (UniqueName: \"kubernetes.io/projected/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-kube-api-access-xj5v2\") pod \"nova-metadata-0\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259316 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-config-data\") pod \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259355 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-config-data\") pod \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259454 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7gv2\" (UniqueName: \"kubernetes.io/projected/a3aafbc7-5b56-4632-97ab-002cc11f4e19-kube-api-access-j7gv2\") pod \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259474 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxjsw\" (UniqueName: \"kubernetes.io/projected/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-kube-api-access-rxjsw\") pod \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259513 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-scripts\") pod \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259554 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-combined-ca-bundle\") pod \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\" (UID: \"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259598 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3aafbc7-5b56-4632-97ab-002cc11f4e19-logs\") pod \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.259615 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-combined-ca-bundle\") pod \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\" (UID: \"a3aafbc7-5b56-4632-97ab-002cc11f4e19\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.260107 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3aafbc7-5b56-4632-97ab-002cc11f4e19-logs" (OuterVolumeSpecName: "logs") pod "a3aafbc7-5b56-4632-97ab-002cc11f4e19" (UID: "a3aafbc7-5b56-4632-97ab-002cc11f4e19"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.260820 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3aafbc7-5b56-4632-97ab-002cc11f4e19-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.264533 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-kube-api-access-rxjsw" (OuterVolumeSpecName: "kube-api-access-rxjsw") pod "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" (UID: "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2"). InnerVolumeSpecName "kube-api-access-rxjsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.265023 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-scripts" (OuterVolumeSpecName: "scripts") pod "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" (UID: "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.273576 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3aafbc7-5b56-4632-97ab-002cc11f4e19-kube-api-access-j7gv2" (OuterVolumeSpecName: "kube-api-access-j7gv2") pod "a3aafbc7-5b56-4632-97ab-002cc11f4e19" (UID: "a3aafbc7-5b56-4632-97ab-002cc11f4e19"). InnerVolumeSpecName "kube-api-access-j7gv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.291698 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3aafbc7-5b56-4632-97ab-002cc11f4e19" (UID: "a3aafbc7-5b56-4632-97ab-002cc11f4e19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.294237 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-config-data" (OuterVolumeSpecName: "config-data") pod "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" (UID: "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.304184 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" (UID: "89fed09c-7c9c-40c8-82e3-dca8f90f0ff2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.306615 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-config-data" (OuterVolumeSpecName: "config-data") pod "a3aafbc7-5b56-4632-97ab-002cc11f4e19" (UID: "a3aafbc7-5b56-4632-97ab-002cc11f4e19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.345963 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362242 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-combined-ca-bundle\") pod \"cbc6695d-ea89-43fe-80e4-b19237977890\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362439 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-config-data\") pod \"cbc6695d-ea89-43fe-80e4-b19237977890\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362534 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lghc8\" (UniqueName: \"kubernetes.io/projected/cbc6695d-ea89-43fe-80e4-b19237977890-kube-api-access-lghc8\") pod \"cbc6695d-ea89-43fe-80e4-b19237977890\" (UID: \"cbc6695d-ea89-43fe-80e4-b19237977890\") " Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362896 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7gv2\" (UniqueName: \"kubernetes.io/projected/a3aafbc7-5b56-4632-97ab-002cc11f4e19-kube-api-access-j7gv2\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362920 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxjsw\" (UniqueName: \"kubernetes.io/projected/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-kube-api-access-rxjsw\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362931 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362941 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362953 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362963 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.362972 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3aafbc7-5b56-4632-97ab-002cc11f4e19-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.367197 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc6695d-ea89-43fe-80e4-b19237977890-kube-api-access-lghc8" (OuterVolumeSpecName: "kube-api-access-lghc8") pod "cbc6695d-ea89-43fe-80e4-b19237977890" (UID: "cbc6695d-ea89-43fe-80e4-b19237977890"). InnerVolumeSpecName "kube-api-access-lghc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.383691 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-config-data" (OuterVolumeSpecName: "config-data") pod "cbc6695d-ea89-43fe-80e4-b19237977890" (UID: "cbc6695d-ea89-43fe-80e4-b19237977890"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.384371 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbc6695d-ea89-43fe-80e4-b19237977890" (UID: "cbc6695d-ea89-43fe-80e4-b19237977890"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.458527 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.463740 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lghc8\" (UniqueName: \"kubernetes.io/projected/cbc6695d-ea89-43fe-80e4-b19237977890-kube-api-access-lghc8\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.463782 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.463792 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc6695d-ea89-43fe-80e4-b19237977890-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.534871 4741 generic.go:334] "Generic (PLEG): container finished" podID="cbc6695d-ea89-43fe-80e4-b19237977890" containerID="e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11" exitCode=0 Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.534929 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbc6695d-ea89-43fe-80e4-b19237977890","Type":"ContainerDied","Data":"e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11"} Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.534955 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbc6695d-ea89-43fe-80e4-b19237977890","Type":"ContainerDied","Data":"b1c0d232b57dee11e15874d09854c48c213d6e9092593cc60c7525e0a8e7f8f1"} Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.534973 4741 scope.go:117] "RemoveContainer" containerID="e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.535062 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.554553 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: E0929 20:41:56.555844 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" containerName="nova-cell1-conductor-db-sync" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.555867 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" containerName="nova-cell1-conductor-db-sync" Sep 29 20:41:56 crc kubenswrapper[4741]: E0929 20:41:56.555901 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc6695d-ea89-43fe-80e4-b19237977890" containerName="nova-scheduler-scheduler" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.555910 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc6695d-ea89-43fe-80e4-b19237977890" containerName="nova-scheduler-scheduler" Sep 29 20:41:56 crc kubenswrapper[4741]: E0929 20:41:56.555929 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-log" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.555937 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-log" Sep 29 20:41:56 crc kubenswrapper[4741]: E0929 20:41:56.555950 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-api" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.555957 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-api" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.556266 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc6695d-ea89-43fe-80e4-b19237977890" containerName="nova-scheduler-scheduler" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.556288 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-api" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.556302 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" containerName="nova-cell1-conductor-db-sync" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.556316 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" containerName="nova-api-log" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.557088 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.560107 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" event={"ID":"89fed09c-7c9c-40c8-82e3-dca8f90f0ff2","Type":"ContainerDied","Data":"9263fd504c23ea2b509a5d3ba65587d021d034a1fb65ec42edd7d9a83a1d899c"} Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.560145 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9263fd504c23ea2b509a5d3ba65587d021d034a1fb65ec42edd7d9a83a1d899c" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.560226 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd5mn" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.564541 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3aafbc7-5b56-4632-97ab-002cc11f4e19","Type":"ContainerDied","Data":"674bc4ea7fd8bfb7f587e6926750c766236218d3da6bfb550e68d43336aa78ea"} Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.564617 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.578289 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5drj2\" (UniqueName: \"kubernetes.io/projected/c989f82b-2539-45be-8138-148762687ef0-kube-api-access-5drj2\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.581087 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.581240 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.596871 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.598431 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.618746 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.623646 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.625894 4741 scope.go:117] "RemoveContainer" containerID="e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11" Sep 29 20:41:56 crc kubenswrapper[4741]: E0929 20:41:56.627276 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11\": container with ID starting with e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11 not found: ID does not exist" containerID="e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.627316 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11"} err="failed to get container status \"e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11\": rpc error: code = NotFound desc = could not find container \"e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11\": container with ID starting with e7ebb15baac2e842e225135117b96b216260962282e066dc9f5de7c42a435d11 not found: ID does not exist" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.627339 4741 scope.go:117] "RemoveContainer" containerID="c309dadd82316dcb4ff62b6317fcdba8e66506f29bcd43681d01c70241afbbbc" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.627554 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.637111 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.656057 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.658228 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.669152 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.684162 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v5l9\" (UniqueName: \"kubernetes.io/projected/c5950e2b-70d4-4352-ad9a-dd53c65839ec-kube-api-access-5v5l9\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.684239 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.684327 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-config-data\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.684372 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5drj2\" (UniqueName: \"kubernetes.io/projected/c989f82b-2539-45be-8138-148762687ef0-kube-api-access-5drj2\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.684411 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.684444 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.691784 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.700047 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.701713 4741 scope.go:117] "RemoveContainer" containerID="7f0f1b826d8acc9a4744a812ca4ae4a1606e3a4b1503ef22b8ae5795805a8ee9" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.701866 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.702944 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.712834 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5drj2\" (UniqueName: \"kubernetes.io/projected/c989f82b-2539-45be-8138-148762687ef0-kube-api-access-5drj2\") pod \"nova-cell1-conductor-0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.713330 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.717156 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.720233 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.752619 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.786262 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59916fd6-9e59-4268-998d-fb8a7fe62249-logs\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.786332 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4fjg\" (UniqueName: \"kubernetes.io/projected/59916fd6-9e59-4268-998d-fb8a7fe62249-kube-api-access-k4fjg\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.786370 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v5l9\" (UniqueName: \"kubernetes.io/projected/c5950e2b-70d4-4352-ad9a-dd53c65839ec-kube-api-access-5v5l9\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.786403 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.786454 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-config-data\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.786498 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-config-data\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.786571 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.789622 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.790839 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-config-data\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.804488 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v5l9\" (UniqueName: \"kubernetes.io/projected/c5950e2b-70d4-4352-ad9a-dd53c65839ec-kube-api-access-5v5l9\") pod \"nova-scheduler-0\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " pod="openstack/nova-scheduler-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.887964 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59916fd6-9e59-4268-998d-fb8a7fe62249-logs\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.888029 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4fjg\" (UniqueName: \"kubernetes.io/projected/59916fd6-9e59-4268-998d-fb8a7fe62249-kube-api-access-k4fjg\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.888066 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.888125 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-config-data\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.888493 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59916fd6-9e59-4268-998d-fb8a7fe62249-logs\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.892715 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-config-data\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.893379 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.903377 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4fjg\" (UniqueName: \"kubernetes.io/projected/59916fd6-9e59-4268-998d-fb8a7fe62249-kube-api-access-k4fjg\") pod \"nova-api-0\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " pod="openstack/nova-api-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.906467 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.947188 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.984259 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:41:56 crc kubenswrapper[4741]: I0929 20:41:56.986954 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.007567 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dffb944f7-sx7bc"] Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.007812 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" podUID="d8f88100-58cc-4534-b686-1656b22af01f" containerName="dnsmasq-dns" containerID="cri-o://c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620" gracePeriod=10 Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.032777 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.119361 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="776b8de5-037e-4cc6-9cb4-d7cefa25f7ad" path="/var/lib/kubelet/pods/776b8de5-037e-4cc6-9cb4-d7cefa25f7ad/volumes" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.120168 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3aafbc7-5b56-4632-97ab-002cc11f4e19" path="/var/lib/kubelet/pods/a3aafbc7-5b56-4632-97ab-002cc11f4e19/volumes" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.120786 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbc6695d-ea89-43fe-80e4-b19237977890" path="/var/lib/kubelet/pods/cbc6695d-ea89-43fe-80e4-b19237977890/volumes" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.497097 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.516666 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.578903 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f","Type":"ContainerStarted","Data":"1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3"} Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.578941 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f","Type":"ContainerStarted","Data":"fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6"} Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.578952 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f","Type":"ContainerStarted","Data":"a725b91c2d54e0ee6e51db3a85a0bd6c7c962e9f26397dec49ce26ae9f0d1472"} Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.581669 4741 generic.go:334] "Generic (PLEG): container finished" podID="d8f88100-58cc-4534-b686-1656b22af01f" containerID="c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620" exitCode=0 Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.581728 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.581738 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" event={"ID":"d8f88100-58cc-4534-b686-1656b22af01f","Type":"ContainerDied","Data":"c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620"} Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.581783 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dffb944f7-sx7bc" event={"ID":"d8f88100-58cc-4534-b686-1656b22af01f","Type":"ContainerDied","Data":"17ca3735e41a71f4254d1f47bfbcd906f88867ce447180447d7d445f9661601e"} Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.581805 4741 scope.go:117] "RemoveContainer" containerID="c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.585684 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c989f82b-2539-45be-8138-148762687ef0","Type":"ContainerStarted","Data":"a90ed34654a93ddd7e71380b6873acb21bce5be97d2e65ec3d5238e611fd6697"} Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.600664 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.608741 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-sb\") pod \"d8f88100-58cc-4534-b686-1656b22af01f\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.608807 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-nb\") pod \"d8f88100-58cc-4534-b686-1656b22af01f\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.608851 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-config\") pod \"d8f88100-58cc-4534-b686-1656b22af01f\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.608935 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-dns-svc\") pod \"d8f88100-58cc-4534-b686-1656b22af01f\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.609092 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8hbk\" (UniqueName: \"kubernetes.io/projected/d8f88100-58cc-4534-b686-1656b22af01f-kube-api-access-m8hbk\") pod \"d8f88100-58cc-4534-b686-1656b22af01f\" (UID: \"d8f88100-58cc-4534-b686-1656b22af01f\") " Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.610240 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.610205549 podStartE2EDuration="2.610205549s" podCreationTimestamp="2025-09-29 20:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:57.604065639 +0000 UTC m=+5559.251854971" watchObservedRunningTime="2025-09-29 20:41:57.610205549 +0000 UTC m=+5559.257994881" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.616096 4741 scope.go:117] "RemoveContainer" containerID="16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.621597 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f88100-58cc-4534-b686-1656b22af01f-kube-api-access-m8hbk" (OuterVolumeSpecName: "kube-api-access-m8hbk") pod "d8f88100-58cc-4534-b686-1656b22af01f" (UID: "d8f88100-58cc-4534-b686-1656b22af01f"). InnerVolumeSpecName "kube-api-access-m8hbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.666602 4741 scope.go:117] "RemoveContainer" containerID="c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.673483 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:41:57 crc kubenswrapper[4741]: E0929 20:41:57.675948 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620\": container with ID starting with c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620 not found: ID does not exist" containerID="c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.676038 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620"} err="failed to get container status \"c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620\": rpc error: code = NotFound desc = could not find container \"c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620\": container with ID starting with c3f535fbd8d10776bcfedaba5b9e67a1ef7bdd9b4a5ec36cd4b27e1baf0f0620 not found: ID does not exist" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.676065 4741 scope.go:117] "RemoveContainer" containerID="16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529" Sep 29 20:41:57 crc kubenswrapper[4741]: E0929 20:41:57.676324 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529\": container with ID starting with 16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529 not found: ID does not exist" containerID="16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.676341 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529"} err="failed to get container status \"16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529\": rpc error: code = NotFound desc = could not find container \"16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529\": container with ID starting with 16f072fdc4d4e9cb6b7691e5575d6bb1c0df89ca8f6a58ae0e5f6fdacea93529 not found: ID does not exist" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.695231 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.705753 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-config" (OuterVolumeSpecName: "config") pod "d8f88100-58cc-4534-b686-1656b22af01f" (UID: "d8f88100-58cc-4534-b686-1656b22af01f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.715295 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8hbk\" (UniqueName: \"kubernetes.io/projected/d8f88100-58cc-4534-b686-1656b22af01f-kube-api-access-m8hbk\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.715354 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.752433 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d8f88100-58cc-4534-b686-1656b22af01f" (UID: "d8f88100-58cc-4534-b686-1656b22af01f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.756871 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d8f88100-58cc-4534-b686-1656b22af01f" (UID: "d8f88100-58cc-4534-b686-1656b22af01f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.757493 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d8f88100-58cc-4534-b686-1656b22af01f" (UID: "d8f88100-58cc-4534-b686-1656b22af01f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.816700 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.817182 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:57 crc kubenswrapper[4741]: I0929 20:41:57.817213 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8f88100-58cc-4534-b686-1656b22af01f-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.051723 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dffb944f7-sx7bc"] Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.059615 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dffb944f7-sx7bc"] Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.595470 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59916fd6-9e59-4268-998d-fb8a7fe62249","Type":"ContainerStarted","Data":"b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc"} Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.595822 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59916fd6-9e59-4268-998d-fb8a7fe62249","Type":"ContainerStarted","Data":"1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d"} Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.595834 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59916fd6-9e59-4268-998d-fb8a7fe62249","Type":"ContainerStarted","Data":"99b6403adf5e9f52e6d9c3368a6c73d7a5bf3b9740d847f00983456a1768fb85"} Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.598450 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5950e2b-70d4-4352-ad9a-dd53c65839ec","Type":"ContainerStarted","Data":"aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66"} Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.598547 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5950e2b-70d4-4352-ad9a-dd53c65839ec","Type":"ContainerStarted","Data":"94a6554cda64705ed738a54b291dfd30d09c9850a671bb7506a465a322136d8e"} Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.600021 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c989f82b-2539-45be-8138-148762687ef0","Type":"ContainerStarted","Data":"a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d"} Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.600150 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.613373 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.6133524809999997 podStartE2EDuration="2.613352481s" podCreationTimestamp="2025-09-29 20:41:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:58.612180364 +0000 UTC m=+5560.259969696" watchObservedRunningTime="2025-09-29 20:41:58.613352481 +0000 UTC m=+5560.261141813" Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.636942 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.63692284 podStartE2EDuration="2.63692284s" podCreationTimestamp="2025-09-29 20:41:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:58.631209313 +0000 UTC m=+5560.278998655" watchObservedRunningTime="2025-09-29 20:41:58.63692284 +0000 UTC m=+5560.284712172" Sep 29 20:41:58 crc kubenswrapper[4741]: I0929 20:41:58.649547 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.649524601 podStartE2EDuration="2.649524601s" podCreationTimestamp="2025-09-29 20:41:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:41:58.644631929 +0000 UTC m=+5560.292421301" watchObservedRunningTime="2025-09-29 20:41:58.649524601 +0000 UTC m=+5560.297313923" Sep 29 20:41:59 crc kubenswrapper[4741]: I0929 20:41:59.102373 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f88100-58cc-4534-b686-1656b22af01f" path="/var/lib/kubelet/pods/d8f88100-58cc-4534-b686-1656b22af01f/volumes" Sep 29 20:42:01 crc kubenswrapper[4741]: I0929 20:42:01.458843 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:42:01 crc kubenswrapper[4741]: I0929 20:42:01.459626 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:42:01 crc kubenswrapper[4741]: I0929 20:42:01.987533 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 20:42:06 crc kubenswrapper[4741]: I0929 20:42:06.459639 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 20:42:06 crc kubenswrapper[4741]: I0929 20:42:06.460512 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 20:42:06 crc kubenswrapper[4741]: I0929 20:42:06.988195 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 20:42:06 crc kubenswrapper[4741]: I0929 20:42:06.993318 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.025496 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.033272 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.033531 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.541669 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.61:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.541809 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.61:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.567127 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2lb6f"] Sep 29 20:42:07 crc kubenswrapper[4741]: E0929 20:42:07.567645 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f88100-58cc-4534-b686-1656b22af01f" containerName="dnsmasq-dns" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.567671 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f88100-58cc-4534-b686-1656b22af01f" containerName="dnsmasq-dns" Sep 29 20:42:07 crc kubenswrapper[4741]: E0929 20:42:07.567704 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f88100-58cc-4534-b686-1656b22af01f" containerName="init" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.567714 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f88100-58cc-4534-b686-1656b22af01f" containerName="init" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.567923 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f88100-58cc-4534-b686-1656b22af01f" containerName="dnsmasq-dns" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.568728 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.572459 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.572736 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.579321 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2lb6f"] Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.711924 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.713077 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-config-data\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.713425 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-scripts\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.713540 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82vlb\" (UniqueName: \"kubernetes.io/projected/5a43ba15-f4b7-45e1-a111-2aecebc65803-kube-api-access-82vlb\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.713639 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.816150 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-scripts\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.816210 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82vlb\" (UniqueName: \"kubernetes.io/projected/5a43ba15-f4b7-45e1-a111-2aecebc65803-kube-api-access-82vlb\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.816306 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.816377 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-config-data\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.823507 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.824207 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-config-data\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.831295 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-scripts\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.845491 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82vlb\" (UniqueName: \"kubernetes.io/projected/5a43ba15-f4b7-45e1-a111-2aecebc65803-kube-api-access-82vlb\") pod \"nova-cell1-cell-mapping-2lb6f\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:07 crc kubenswrapper[4741]: I0929 20:42:07.905757 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:08 crc kubenswrapper[4741]: I0929 20:42:08.119590 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.64:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:08 crc kubenswrapper[4741]: I0929 20:42:08.119628 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.64:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:08 crc kubenswrapper[4741]: I0929 20:42:08.436333 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2lb6f"] Sep 29 20:42:08 crc kubenswrapper[4741]: I0929 20:42:08.693709 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2lb6f" event={"ID":"5a43ba15-f4b7-45e1-a111-2aecebc65803","Type":"ContainerStarted","Data":"15bdcf1fd3f5ad55854ac6d22543b49505c17938181d65937cc7cc0c6a38d6d0"} Sep 29 20:42:08 crc kubenswrapper[4741]: I0929 20:42:08.694135 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2lb6f" event={"ID":"5a43ba15-f4b7-45e1-a111-2aecebc65803","Type":"ContainerStarted","Data":"269f0e7fc91f1e8291957b1524ebf5e3695cbbf20250c895f095ec7a3ba27d4a"} Sep 29 20:42:08 crc kubenswrapper[4741]: I0929 20:42:08.713401 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2lb6f" podStartSLOduration=1.713359268 podStartE2EDuration="1.713359268s" podCreationTimestamp="2025-09-29 20:42:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:42:08.709113628 +0000 UTC m=+5570.356902970" watchObservedRunningTime="2025-09-29 20:42:08.713359268 +0000 UTC m=+5570.361148600" Sep 29 20:42:13 crc kubenswrapper[4741]: I0929 20:42:13.739832 4741 generic.go:334] "Generic (PLEG): container finished" podID="5a43ba15-f4b7-45e1-a111-2aecebc65803" containerID="15bdcf1fd3f5ad55854ac6d22543b49505c17938181d65937cc7cc0c6a38d6d0" exitCode=0 Sep 29 20:42:13 crc kubenswrapper[4741]: I0929 20:42:13.739906 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2lb6f" event={"ID":"5a43ba15-f4b7-45e1-a111-2aecebc65803","Type":"ContainerDied","Data":"15bdcf1fd3f5ad55854ac6d22543b49505c17938181d65937cc7cc0c6a38d6d0"} Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.124327 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.263103 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-config-data\") pod \"5a43ba15-f4b7-45e1-a111-2aecebc65803\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.263286 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-scripts\") pod \"5a43ba15-f4b7-45e1-a111-2aecebc65803\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.263405 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82vlb\" (UniqueName: \"kubernetes.io/projected/5a43ba15-f4b7-45e1-a111-2aecebc65803-kube-api-access-82vlb\") pod \"5a43ba15-f4b7-45e1-a111-2aecebc65803\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.263488 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-combined-ca-bundle\") pod \"5a43ba15-f4b7-45e1-a111-2aecebc65803\" (UID: \"5a43ba15-f4b7-45e1-a111-2aecebc65803\") " Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.270428 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a43ba15-f4b7-45e1-a111-2aecebc65803-kube-api-access-82vlb" (OuterVolumeSpecName: "kube-api-access-82vlb") pod "5a43ba15-f4b7-45e1-a111-2aecebc65803" (UID: "5a43ba15-f4b7-45e1-a111-2aecebc65803"). InnerVolumeSpecName "kube-api-access-82vlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.271260 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-scripts" (OuterVolumeSpecName: "scripts") pod "5a43ba15-f4b7-45e1-a111-2aecebc65803" (UID: "5a43ba15-f4b7-45e1-a111-2aecebc65803"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.291162 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a43ba15-f4b7-45e1-a111-2aecebc65803" (UID: "5a43ba15-f4b7-45e1-a111-2aecebc65803"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.298662 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-config-data" (OuterVolumeSpecName: "config-data") pod "5a43ba15-f4b7-45e1-a111-2aecebc65803" (UID: "5a43ba15-f4b7-45e1-a111-2aecebc65803"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.365242 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.365286 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82vlb\" (UniqueName: \"kubernetes.io/projected/5a43ba15-f4b7-45e1-a111-2aecebc65803-kube-api-access-82vlb\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.365302 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.365314 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a43ba15-f4b7-45e1-a111-2aecebc65803-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.763244 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2lb6f" event={"ID":"5a43ba15-f4b7-45e1-a111-2aecebc65803","Type":"ContainerDied","Data":"269f0e7fc91f1e8291957b1524ebf5e3695cbbf20250c895f095ec7a3ba27d4a"} Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.763602 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="269f0e7fc91f1e8291957b1524ebf5e3695cbbf20250c895f095ec7a3ba27d4a" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.763310 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2lb6f" Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.930198 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.930447 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-log" containerID="cri-o://1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d" gracePeriod=30 Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.930577 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-api" containerID="cri-o://b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc" gracePeriod=30 Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.952397 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.952639 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c5950e2b-70d4-4352-ad9a-dd53c65839ec" containerName="nova-scheduler-scheduler" containerID="cri-o://aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" gracePeriod=30 Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.964145 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.964568 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-log" containerID="cri-o://fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6" gracePeriod=30 Sep 29 20:42:15 crc kubenswrapper[4741]: I0929 20:42:15.964673 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-metadata" containerID="cri-o://1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3" gracePeriod=30 Sep 29 20:42:16 crc kubenswrapper[4741]: I0929 20:42:16.772308 4741 generic.go:334] "Generic (PLEG): container finished" podID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerID="1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d" exitCode=143 Sep 29 20:42:16 crc kubenswrapper[4741]: I0929 20:42:16.772361 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59916fd6-9e59-4268-998d-fb8a7fe62249","Type":"ContainerDied","Data":"1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d"} Sep 29 20:42:16 crc kubenswrapper[4741]: I0929 20:42:16.774855 4741 generic.go:334] "Generic (PLEG): container finished" podID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerID="fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6" exitCode=143 Sep 29 20:42:16 crc kubenswrapper[4741]: I0929 20:42:16.774880 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f","Type":"ContainerDied","Data":"fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6"} Sep 29 20:42:16 crc kubenswrapper[4741]: E0929 20:42:16.991383 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 20:42:16 crc kubenswrapper[4741]: E0929 20:42:16.992777 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 20:42:16 crc kubenswrapper[4741]: E0929 20:42:16.994831 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 20:42:16 crc kubenswrapper[4741]: E0929 20:42:16.994909 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c5950e2b-70d4-4352-ad9a-dd53c65839ec" containerName="nova-scheduler-scheduler" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.338711 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.421536 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.435054 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-combined-ca-bundle\") pod \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.435145 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v5l9\" (UniqueName: \"kubernetes.io/projected/c5950e2b-70d4-4352-ad9a-dd53c65839ec-kube-api-access-5v5l9\") pod \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.435216 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-config-data\") pod \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\" (UID: \"c5950e2b-70d4-4352-ad9a-dd53c65839ec\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.443510 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5950e2b-70d4-4352-ad9a-dd53c65839ec-kube-api-access-5v5l9" (OuterVolumeSpecName: "kube-api-access-5v5l9") pod "c5950e2b-70d4-4352-ad9a-dd53c65839ec" (UID: "c5950e2b-70d4-4352-ad9a-dd53c65839ec"). InnerVolumeSpecName "kube-api-access-5v5l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.485038 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5950e2b-70d4-4352-ad9a-dd53c65839ec" (UID: "c5950e2b-70d4-4352-ad9a-dd53c65839ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.503907 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-config-data" (OuterVolumeSpecName: "config-data") pod "c5950e2b-70d4-4352-ad9a-dd53c65839ec" (UID: "c5950e2b-70d4-4352-ad9a-dd53c65839ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.514101 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.538802 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-combined-ca-bundle\") pod \"59916fd6-9e59-4268-998d-fb8a7fe62249\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.539518 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4fjg\" (UniqueName: \"kubernetes.io/projected/59916fd6-9e59-4268-998d-fb8a7fe62249-kube-api-access-k4fjg\") pod \"59916fd6-9e59-4268-998d-fb8a7fe62249\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.539552 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59916fd6-9e59-4268-998d-fb8a7fe62249-logs\") pod \"59916fd6-9e59-4268-998d-fb8a7fe62249\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.539617 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-config-data\") pod \"59916fd6-9e59-4268-998d-fb8a7fe62249\" (UID: \"59916fd6-9e59-4268-998d-fb8a7fe62249\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.539961 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.539981 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v5l9\" (UniqueName: \"kubernetes.io/projected/c5950e2b-70d4-4352-ad9a-dd53c65839ec-kube-api-access-5v5l9\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.539992 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5950e2b-70d4-4352-ad9a-dd53c65839ec-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.540970 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59916fd6-9e59-4268-998d-fb8a7fe62249-logs" (OuterVolumeSpecName: "logs") pod "59916fd6-9e59-4268-998d-fb8a7fe62249" (UID: "59916fd6-9e59-4268-998d-fb8a7fe62249"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.545524 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59916fd6-9e59-4268-998d-fb8a7fe62249-kube-api-access-k4fjg" (OuterVolumeSpecName: "kube-api-access-k4fjg") pod "59916fd6-9e59-4268-998d-fb8a7fe62249" (UID: "59916fd6-9e59-4268-998d-fb8a7fe62249"). InnerVolumeSpecName "kube-api-access-k4fjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.562549 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59916fd6-9e59-4268-998d-fb8a7fe62249" (UID: "59916fd6-9e59-4268-998d-fb8a7fe62249"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.567877 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-config-data" (OuterVolumeSpecName: "config-data") pod "59916fd6-9e59-4268-998d-fb8a7fe62249" (UID: "59916fd6-9e59-4268-998d-fb8a7fe62249"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.641467 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-combined-ca-bundle\") pod \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.641528 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-logs\") pod \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.641549 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj5v2\" (UniqueName: \"kubernetes.io/projected/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-kube-api-access-xj5v2\") pod \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.641637 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-config-data\") pod \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\" (UID: \"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f\") " Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.642014 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4fjg\" (UniqueName: \"kubernetes.io/projected/59916fd6-9e59-4268-998d-fb8a7fe62249-kube-api-access-k4fjg\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.642030 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59916fd6-9e59-4268-998d-fb8a7fe62249-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.642040 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.642048 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59916fd6-9e59-4268-998d-fb8a7fe62249-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.643262 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-logs" (OuterVolumeSpecName: "logs") pod "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" (UID: "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.645655 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-kube-api-access-xj5v2" (OuterVolumeSpecName: "kube-api-access-xj5v2") pod "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" (UID: "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f"). InnerVolumeSpecName "kube-api-access-xj5v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.664544 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-config-data" (OuterVolumeSpecName: "config-data") pod "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" (UID: "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.671405 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" (UID: "cc985ca1-ce40-4f74-9f17-9c5f1fe0163f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.743626 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.743669 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.743679 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj5v2\" (UniqueName: \"kubernetes.io/projected/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-kube-api-access-xj5v2\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.743690 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.802156 4741 generic.go:334] "Generic (PLEG): container finished" podID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerID="b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc" exitCode=0 Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.802225 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59916fd6-9e59-4268-998d-fb8a7fe62249","Type":"ContainerDied","Data":"b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc"} Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.802254 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59916fd6-9e59-4268-998d-fb8a7fe62249","Type":"ContainerDied","Data":"99b6403adf5e9f52e6d9c3368a6c73d7a5bf3b9740d847f00983456a1768fb85"} Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.802272 4741 scope.go:117] "RemoveContainer" containerID="b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.802296 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.805235 4741 generic.go:334] "Generic (PLEG): container finished" podID="c5950e2b-70d4-4352-ad9a-dd53c65839ec" containerID="aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" exitCode=0 Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.805294 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5950e2b-70d4-4352-ad9a-dd53c65839ec","Type":"ContainerDied","Data":"aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66"} Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.805314 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5950e2b-70d4-4352-ad9a-dd53c65839ec","Type":"ContainerDied","Data":"94a6554cda64705ed738a54b291dfd30d09c9850a671bb7506a465a322136d8e"} Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.805329 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.808611 4741 generic.go:334] "Generic (PLEG): container finished" podID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerID="1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3" exitCode=0 Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.808676 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f","Type":"ContainerDied","Data":"1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3"} Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.808713 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc985ca1-ce40-4f74-9f17-9c5f1fe0163f","Type":"ContainerDied","Data":"a725b91c2d54e0ee6e51db3a85a0bd6c7c962e9f26397dec49ce26ae9f0d1472"} Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.808797 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.839890 4741 scope.go:117] "RemoveContainer" containerID="1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.862681 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.887741 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.904514 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.905573 4741 scope.go:117] "RemoveContainer" containerID="b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.911934 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc\": container with ID starting with b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc not found: ID does not exist" containerID="b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.911982 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc"} err="failed to get container status \"b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc\": rpc error: code = NotFound desc = could not find container \"b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc\": container with ID starting with b2b9940c63e4f5dc9e467e427f6416d6322ce9b55e2d95c265cc1125172fccdc not found: ID does not exist" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.912053 4741 scope.go:117] "RemoveContainer" containerID="1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.916612 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d\": container with ID starting with 1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d not found: ID does not exist" containerID="1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.916666 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d"} err="failed to get container status \"1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d\": rpc error: code = NotFound desc = could not find container \"1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d\": container with ID starting with 1902403b287a16c5b46aa78dacaf98e0e9838df3555dd7c9b898500c8def788d not found: ID does not exist" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.916693 4741 scope.go:117] "RemoveContainer" containerID="aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.919108 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931011 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.931535 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a43ba15-f4b7-45e1-a111-2aecebc65803" containerName="nova-manage" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931554 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a43ba15-f4b7-45e1-a111-2aecebc65803" containerName="nova-manage" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.931584 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-log" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931593 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-log" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.931615 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-api" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931624 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-api" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.931643 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-log" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931651 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-log" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.931672 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5950e2b-70d4-4352-ad9a-dd53c65839ec" containerName="nova-scheduler-scheduler" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931678 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5950e2b-70d4-4352-ad9a-dd53c65839ec" containerName="nova-scheduler-scheduler" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.931692 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-metadata" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931701 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-metadata" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931882 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-api" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931896 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-metadata" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931904 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5950e2b-70d4-4352-ad9a-dd53c65839ec" containerName="nova-scheduler-scheduler" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931913 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a43ba15-f4b7-45e1-a111-2aecebc65803" containerName="nova-manage" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931923 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" containerName="nova-metadata-log" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.931936 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" containerName="nova-api-log" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.932904 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.935573 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.937318 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.946875 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.948348 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.952014 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.957174 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.963683 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.964052 4741 scope.go:117] "RemoveContainer" containerID="aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" Sep 29 20:42:19 crc kubenswrapper[4741]: E0929 20:42:19.964595 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66\": container with ID starting with aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66 not found: ID does not exist" containerID="aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.964623 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66"} err="failed to get container status \"aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66\": rpc error: code = NotFound desc = could not find container \"aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66\": container with ID starting with aab37492d9e94d8b843f8f22914de73d144aa5e4b90be250fb6d50995dfaed66 not found: ID does not exist" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.964669 4741 scope.go:117] "RemoveContainer" containerID="1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.975330 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.979834 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.981658 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.983768 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.987012 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:42:19 crc kubenswrapper[4741]: I0929 20:42:19.989734 4741 scope.go:117] "RemoveContainer" containerID="fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.009320 4741 scope.go:117] "RemoveContainer" containerID="1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3" Sep 29 20:42:20 crc kubenswrapper[4741]: E0929 20:42:20.009697 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3\": container with ID starting with 1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3 not found: ID does not exist" containerID="1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.009728 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3"} err="failed to get container status \"1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3\": rpc error: code = NotFound desc = could not find container \"1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3\": container with ID starting with 1a994e67885f42b3dae906e67991a271ccc5aa7340e8a40808125d258f4212f3 not found: ID does not exist" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.009750 4741 scope.go:117] "RemoveContainer" containerID="fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6" Sep 29 20:42:20 crc kubenswrapper[4741]: E0929 20:42:20.009964 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6\": container with ID starting with fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6 not found: ID does not exist" containerID="fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.009986 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6"} err="failed to get container status \"fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6\": rpc error: code = NotFound desc = could not find container \"fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6\": container with ID starting with fba626ff0e8222e54eafb246dd063067c97f45eb6cd9943b4a40a936a7c46fa6 not found: ID does not exist" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049206 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-config-data\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049314 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff279576-11ac-4c99-9b06-934b9c6e2257-logs\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049341 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f20bc6-75ac-42bf-957b-37b925499cd2-logs\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049422 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049446 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd6pw\" (UniqueName: \"kubernetes.io/projected/30f20bc6-75ac-42bf-957b-37b925499cd2-kube-api-access-gd6pw\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049476 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049520 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7bnw\" (UniqueName: \"kubernetes.io/projected/d813c027-ccbb-41f2-8ae1-4e547b1ad080-kube-api-access-l7bnw\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049633 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpzh8\" (UniqueName: \"kubernetes.io/projected/ff279576-11ac-4c99-9b06-934b9c6e2257-kube-api-access-zpzh8\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049741 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049811 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-config-data\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.049951 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-config-data\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152015 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-config-data\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152105 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-config-data\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152151 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff279576-11ac-4c99-9b06-934b9c6e2257-logs\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152191 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f20bc6-75ac-42bf-957b-37b925499cd2-logs\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152220 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152261 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd6pw\" (UniqueName: \"kubernetes.io/projected/30f20bc6-75ac-42bf-957b-37b925499cd2-kube-api-access-gd6pw\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152290 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152333 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7bnw\" (UniqueName: \"kubernetes.io/projected/d813c027-ccbb-41f2-8ae1-4e547b1ad080-kube-api-access-l7bnw\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152352 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpzh8\" (UniqueName: \"kubernetes.io/projected/ff279576-11ac-4c99-9b06-934b9c6e2257-kube-api-access-zpzh8\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152379 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152430 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-config-data\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152714 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f20bc6-75ac-42bf-957b-37b925499cd2-logs\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.152773 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff279576-11ac-4c99-9b06-934b9c6e2257-logs\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.156419 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.157847 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.158855 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-config-data\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.159100 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-config-data\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.160553 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-config-data\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.160765 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.174500 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7bnw\" (UniqueName: \"kubernetes.io/projected/d813c027-ccbb-41f2-8ae1-4e547b1ad080-kube-api-access-l7bnw\") pod \"nova-scheduler-0\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.176866 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpzh8\" (UniqueName: \"kubernetes.io/projected/ff279576-11ac-4c99-9b06-934b9c6e2257-kube-api-access-zpzh8\") pod \"nova-api-0\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.177177 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd6pw\" (UniqueName: \"kubernetes.io/projected/30f20bc6-75ac-42bf-957b-37b925499cd2-kube-api-access-gd6pw\") pod \"nova-metadata-0\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.256117 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.273630 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.295567 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.702032 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:42:20 crc kubenswrapper[4741]: W0929 20:42:20.704232 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd813c027_ccbb_41f2_8ae1_4e547b1ad080.slice/crio-9ad0edd31fcefc4c79019b0b6c8175a12217bdb1ceaed1843224e5259160ac96 WatchSource:0}: Error finding container 9ad0edd31fcefc4c79019b0b6c8175a12217bdb1ceaed1843224e5259160ac96: Status 404 returned error can't find the container with id 9ad0edd31fcefc4c79019b0b6c8175a12217bdb1ceaed1843224e5259160ac96 Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.766584 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.773929 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:42:20 crc kubenswrapper[4741]: W0929 20:42:20.779919 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff279576_11ac_4c99_9b06_934b9c6e2257.slice/crio-56a31fc65c75aa652940183723d69dc40ad61e6279d052251f3aef0d7a6dfc0b WatchSource:0}: Error finding container 56a31fc65c75aa652940183723d69dc40ad61e6279d052251f3aef0d7a6dfc0b: Status 404 returned error can't find the container with id 56a31fc65c75aa652940183723d69dc40ad61e6279d052251f3aef0d7a6dfc0b Sep 29 20:42:20 crc kubenswrapper[4741]: W0929 20:42:20.781265 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30f20bc6_75ac_42bf_957b_37b925499cd2.slice/crio-88a40d8642f6f3e121141148e11b223d5e5a41c0d235d3d4b2bf0beff687c2c6 WatchSource:0}: Error finding container 88a40d8642f6f3e121141148e11b223d5e5a41c0d235d3d4b2bf0beff687c2c6: Status 404 returned error can't find the container with id 88a40d8642f6f3e121141148e11b223d5e5a41c0d235d3d4b2bf0beff687c2c6 Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.826220 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d813c027-ccbb-41f2-8ae1-4e547b1ad080","Type":"ContainerStarted","Data":"9ad0edd31fcefc4c79019b0b6c8175a12217bdb1ceaed1843224e5259160ac96"} Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.828321 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff279576-11ac-4c99-9b06-934b9c6e2257","Type":"ContainerStarted","Data":"56a31fc65c75aa652940183723d69dc40ad61e6279d052251f3aef0d7a6dfc0b"} Sep 29 20:42:20 crc kubenswrapper[4741]: I0929 20:42:20.829324 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"30f20bc6-75ac-42bf-957b-37b925499cd2","Type":"ContainerStarted","Data":"88a40d8642f6f3e121141148e11b223d5e5a41c0d235d3d4b2bf0beff687c2c6"} Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.098488 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59916fd6-9e59-4268-998d-fb8a7fe62249" path="/var/lib/kubelet/pods/59916fd6-9e59-4268-998d-fb8a7fe62249/volumes" Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.099040 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5950e2b-70d4-4352-ad9a-dd53c65839ec" path="/var/lib/kubelet/pods/c5950e2b-70d4-4352-ad9a-dd53c65839ec/volumes" Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.099600 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc985ca1-ce40-4f74-9f17-9c5f1fe0163f" path="/var/lib/kubelet/pods/cc985ca1-ce40-4f74-9f17-9c5f1fe0163f/volumes" Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.869563 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d813c027-ccbb-41f2-8ae1-4e547b1ad080","Type":"ContainerStarted","Data":"8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d"} Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.871340 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff279576-11ac-4c99-9b06-934b9c6e2257","Type":"ContainerStarted","Data":"3d842fabb3d53f20f0e136863c5ebb114fb5e99f3814b59d00241ba9fe2fffba"} Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.871364 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff279576-11ac-4c99-9b06-934b9c6e2257","Type":"ContainerStarted","Data":"e1e215ca44b835d24fe577aa39c784d4c217a95d0c263059037cb67e6daa7178"} Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.873736 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"30f20bc6-75ac-42bf-957b-37b925499cd2","Type":"ContainerStarted","Data":"96dc27c4a2c3bd098e5de6a697b6cd4762f6e2da63b55b540d548e98b85ecca5"} Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.873762 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"30f20bc6-75ac-42bf-957b-37b925499cd2","Type":"ContainerStarted","Data":"20bc3f9ead0984f7f0fb2ab8727a8909152dc8949fa6c4c13a884b281865cc78"} Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.894803 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8947873509999997 podStartE2EDuration="2.894787351s" podCreationTimestamp="2025-09-29 20:42:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:42:21.888759544 +0000 UTC m=+5583.536548876" watchObservedRunningTime="2025-09-29 20:42:21.894787351 +0000 UTC m=+5583.542576683" Sep 29 20:42:21 crc kubenswrapper[4741]: I0929 20:42:21.913596 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.913575612 podStartE2EDuration="2.913575612s" podCreationTimestamp="2025-09-29 20:42:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:42:21.905783411 +0000 UTC m=+5583.553572743" watchObservedRunningTime="2025-09-29 20:42:21.913575612 +0000 UTC m=+5583.561364944" Sep 29 20:42:25 crc kubenswrapper[4741]: I0929 20:42:25.274209 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 20:42:25 crc kubenswrapper[4741]: I0929 20:42:25.296496 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:42:25 crc kubenswrapper[4741]: I0929 20:42:25.296578 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.257228 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.257889 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.274598 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.295796 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.295845 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.309126 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.343239 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=11.343216935 podStartE2EDuration="11.343216935s" podCreationTimestamp="2025-09-29 20:42:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:42:21.929897487 +0000 UTC m=+5583.577686819" watchObservedRunningTime="2025-09-29 20:42:30.343216935 +0000 UTC m=+5591.991006257" Sep 29 20:42:30 crc kubenswrapper[4741]: I0929 20:42:30.985989 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 20:42:31 crc kubenswrapper[4741]: I0929 20:42:31.421604 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:31 crc kubenswrapper[4741]: I0929 20:42:31.421665 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.66:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:31 crc kubenswrapper[4741]: I0929 20:42:31.421953 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.66:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:31 crc kubenswrapper[4741]: I0929 20:42:31.422011 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.260007 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.260648 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.260932 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.261050 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.272108 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.272209 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.299856 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.304831 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.312861 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.451484 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d8cc9c9c-8vpqg"] Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.453391 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.458519 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d8cc9c9c-8vpqg"] Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.545572 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-config\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.545618 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-nb\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.545651 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-dns-svc\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.545845 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5w5b\" (UniqueName: \"kubernetes.io/projected/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-kube-api-access-l5w5b\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.545982 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-sb\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.647833 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5w5b\" (UniqueName: \"kubernetes.io/projected/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-kube-api-access-l5w5b\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.647922 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-sb\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.648049 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-config\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.648075 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-nb\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.648110 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-dns-svc\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.648998 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-sb\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.649041 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-config\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.649141 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-dns-svc\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.649360 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-nb\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.667203 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5w5b\" (UniqueName: \"kubernetes.io/projected/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-kube-api-access-l5w5b\") pod \"dnsmasq-dns-68d8cc9c9c-8vpqg\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:40 crc kubenswrapper[4741]: I0929 20:42:40.811202 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:41 crc kubenswrapper[4741]: I0929 20:42:41.066152 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 20:42:41 crc kubenswrapper[4741]: I0929 20:42:41.284457 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d8cc9c9c-8vpqg"] Sep 29 20:42:42 crc kubenswrapper[4741]: I0929 20:42:42.067287 4741 generic.go:334] "Generic (PLEG): container finished" podID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerID="b941aef73311cb3c9242cfc4a773bef4fc1a78d72397c2623c5b2398840b5989" exitCode=0 Sep 29 20:42:42 crc kubenswrapper[4741]: I0929 20:42:42.068476 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" event={"ID":"d20b7eec-51aa-4fbf-a6ec-58f05a97e540","Type":"ContainerDied","Data":"b941aef73311cb3c9242cfc4a773bef4fc1a78d72397c2623c5b2398840b5989"} Sep 29 20:42:42 crc kubenswrapper[4741]: I0929 20:42:42.068526 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" event={"ID":"d20b7eec-51aa-4fbf-a6ec-58f05a97e540","Type":"ContainerStarted","Data":"f8e412083e0526090bd6dac805d4db360fb1a2e6019a86bd3f8ca7307552ab31"} Sep 29 20:42:43 crc kubenswrapper[4741]: I0929 20:42:43.078172 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" event={"ID":"d20b7eec-51aa-4fbf-a6ec-58f05a97e540","Type":"ContainerStarted","Data":"f9e835d64da2e05f43f3fd7a734d412f4da9ae5eac2a36ff86c9341bc1ad0c78"} Sep 29 20:42:43 crc kubenswrapper[4741]: I0929 20:42:43.109746 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" podStartSLOduration=3.109719461 podStartE2EDuration="3.109719461s" podCreationTimestamp="2025-09-29 20:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:42:43.099336509 +0000 UTC m=+5604.747125841" watchObservedRunningTime="2025-09-29 20:42:43.109719461 +0000 UTC m=+5604.757508793" Sep 29 20:42:44 crc kubenswrapper[4741]: I0929 20:42:44.085004 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:50 crc kubenswrapper[4741]: I0929 20:42:50.813447 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:42:50 crc kubenswrapper[4741]: I0929 20:42:50.892476 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d56b878c9-4pvdq"] Sep 29 20:42:50 crc kubenswrapper[4741]: I0929 20:42:50.892796 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" podUID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerName="dnsmasq-dns" containerID="cri-o://91d3646c187680b47e8a22e392a96da2a25cdea71c527b0f773fed6761b5a71a" gracePeriod=10 Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.175998 4741 generic.go:334] "Generic (PLEG): container finished" podID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerID="91d3646c187680b47e8a22e392a96da2a25cdea71c527b0f773fed6761b5a71a" exitCode=0 Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.176064 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" event={"ID":"eab69ab2-6d74-4862-9a59-42fa5c09ccaa","Type":"ContainerDied","Data":"91d3646c187680b47e8a22e392a96da2a25cdea71c527b0f773fed6761b5a71a"} Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.446381 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.525075 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-nb\") pod \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.525229 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgf4s\" (UniqueName: \"kubernetes.io/projected/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-kube-api-access-qgf4s\") pod \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.525285 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-dns-svc\") pod \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.525345 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-config\") pod \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.525476 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-sb\") pod \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\" (UID: \"eab69ab2-6d74-4862-9a59-42fa5c09ccaa\") " Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.531181 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-kube-api-access-qgf4s" (OuterVolumeSpecName: "kube-api-access-qgf4s") pod "eab69ab2-6d74-4862-9a59-42fa5c09ccaa" (UID: "eab69ab2-6d74-4862-9a59-42fa5c09ccaa"). InnerVolumeSpecName "kube-api-access-qgf4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.569502 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-config" (OuterVolumeSpecName: "config") pod "eab69ab2-6d74-4862-9a59-42fa5c09ccaa" (UID: "eab69ab2-6d74-4862-9a59-42fa5c09ccaa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.569799 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eab69ab2-6d74-4862-9a59-42fa5c09ccaa" (UID: "eab69ab2-6d74-4862-9a59-42fa5c09ccaa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.570204 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eab69ab2-6d74-4862-9a59-42fa5c09ccaa" (UID: "eab69ab2-6d74-4862-9a59-42fa5c09ccaa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.574346 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eab69ab2-6d74-4862-9a59-42fa5c09ccaa" (UID: "eab69ab2-6d74-4862-9a59-42fa5c09ccaa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.627762 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.627950 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.628012 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgf4s\" (UniqueName: \"kubernetes.io/projected/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-kube-api-access-qgf4s\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.628068 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:51 crc kubenswrapper[4741]: I0929 20:42:51.628161 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab69ab2-6d74-4862-9a59-42fa5c09ccaa-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:52 crc kubenswrapper[4741]: I0929 20:42:52.191506 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" event={"ID":"eab69ab2-6d74-4862-9a59-42fa5c09ccaa","Type":"ContainerDied","Data":"0f4092149aad532f4c8df961b8e8082c8def6078230a2ea1be021430931588ea"} Sep 29 20:42:52 crc kubenswrapper[4741]: I0929 20:42:52.191600 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d56b878c9-4pvdq" Sep 29 20:42:52 crc kubenswrapper[4741]: I0929 20:42:52.192288 4741 scope.go:117] "RemoveContainer" containerID="91d3646c187680b47e8a22e392a96da2a25cdea71c527b0f773fed6761b5a71a" Sep 29 20:42:52 crc kubenswrapper[4741]: I0929 20:42:52.219507 4741 scope.go:117] "RemoveContainer" containerID="119d985cb03d25915f1ec0f93863619c09e7b456672ca6d486d3fefdac475d82" Sep 29 20:42:52 crc kubenswrapper[4741]: I0929 20:42:52.255690 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d56b878c9-4pvdq"] Sep 29 20:42:52 crc kubenswrapper[4741]: I0929 20:42:52.264768 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d56b878c9-4pvdq"] Sep 29 20:42:53 crc kubenswrapper[4741]: I0929 20:42:53.102834 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" path="/var/lib/kubelet/pods/eab69ab2-6d74-4862-9a59-42fa5c09ccaa/volumes" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.187214 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-dc4f2"] Sep 29 20:42:54 crc kubenswrapper[4741]: E0929 20:42:54.187957 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerName="dnsmasq-dns" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.187971 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerName="dnsmasq-dns" Sep 29 20:42:54 crc kubenswrapper[4741]: E0929 20:42:54.187986 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerName="init" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.187993 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerName="init" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.188181 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab69ab2-6d74-4862-9a59-42fa5c09ccaa" containerName="dnsmasq-dns" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.188900 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dc4f2" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.196091 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-dc4f2"] Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.295334 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgxr6\" (UniqueName: \"kubernetes.io/projected/e93df880-9269-48e4-9815-bbd513579722-kube-api-access-zgxr6\") pod \"cinder-db-create-dc4f2\" (UID: \"e93df880-9269-48e4-9815-bbd513579722\") " pod="openstack/cinder-db-create-dc4f2" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.396830 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgxr6\" (UniqueName: \"kubernetes.io/projected/e93df880-9269-48e4-9815-bbd513579722-kube-api-access-zgxr6\") pod \"cinder-db-create-dc4f2\" (UID: \"e93df880-9269-48e4-9815-bbd513579722\") " pod="openstack/cinder-db-create-dc4f2" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.415531 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgxr6\" (UniqueName: \"kubernetes.io/projected/e93df880-9269-48e4-9815-bbd513579722-kube-api-access-zgxr6\") pod \"cinder-db-create-dc4f2\" (UID: \"e93df880-9269-48e4-9815-bbd513579722\") " pod="openstack/cinder-db-create-dc4f2" Sep 29 20:42:54 crc kubenswrapper[4741]: I0929 20:42:54.514815 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dc4f2" Sep 29 20:42:55 crc kubenswrapper[4741]: W0929 20:42:55.094093 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode93df880_9269_48e4_9815_bbd513579722.slice/crio-7020456ac76fc76619766e2cf63e3e6f65fbc223f39f2e5c79d3c1d34bec8063 WatchSource:0}: Error finding container 7020456ac76fc76619766e2cf63e3e6f65fbc223f39f2e5c79d3c1d34bec8063: Status 404 returned error can't find the container with id 7020456ac76fc76619766e2cf63e3e6f65fbc223f39f2e5c79d3c1d34bec8063 Sep 29 20:42:55 crc kubenswrapper[4741]: I0929 20:42:55.101828 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-dc4f2"] Sep 29 20:42:55 crc kubenswrapper[4741]: I0929 20:42:55.222186 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dc4f2" event={"ID":"e93df880-9269-48e4-9815-bbd513579722","Type":"ContainerStarted","Data":"7020456ac76fc76619766e2cf63e3e6f65fbc223f39f2e5c79d3c1d34bec8063"} Sep 29 20:42:56 crc kubenswrapper[4741]: I0929 20:42:56.234235 4741 generic.go:334] "Generic (PLEG): container finished" podID="e93df880-9269-48e4-9815-bbd513579722" containerID="fbf8f9b90c9606b691caf816febd66a5f65f40a818d40e4b04d6ba134fc57f43" exitCode=0 Sep 29 20:42:56 crc kubenswrapper[4741]: I0929 20:42:56.234377 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dc4f2" event={"ID":"e93df880-9269-48e4-9815-bbd513579722","Type":"ContainerDied","Data":"fbf8f9b90c9606b691caf816febd66a5f65f40a818d40e4b04d6ba134fc57f43"} Sep 29 20:42:57 crc kubenswrapper[4741]: I0929 20:42:57.576981 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dc4f2" Sep 29 20:42:57 crc kubenswrapper[4741]: I0929 20:42:57.667466 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgxr6\" (UniqueName: \"kubernetes.io/projected/e93df880-9269-48e4-9815-bbd513579722-kube-api-access-zgxr6\") pod \"e93df880-9269-48e4-9815-bbd513579722\" (UID: \"e93df880-9269-48e4-9815-bbd513579722\") " Sep 29 20:42:57 crc kubenswrapper[4741]: I0929 20:42:57.675855 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93df880-9269-48e4-9815-bbd513579722-kube-api-access-zgxr6" (OuterVolumeSpecName: "kube-api-access-zgxr6") pod "e93df880-9269-48e4-9815-bbd513579722" (UID: "e93df880-9269-48e4-9815-bbd513579722"). InnerVolumeSpecName "kube-api-access-zgxr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:42:57 crc kubenswrapper[4741]: I0929 20:42:57.770005 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgxr6\" (UniqueName: \"kubernetes.io/projected/e93df880-9269-48e4-9815-bbd513579722-kube-api-access-zgxr6\") on node \"crc\" DevicePath \"\"" Sep 29 20:42:58 crc kubenswrapper[4741]: I0929 20:42:58.259686 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dc4f2" event={"ID":"e93df880-9269-48e4-9815-bbd513579722","Type":"ContainerDied","Data":"7020456ac76fc76619766e2cf63e3e6f65fbc223f39f2e5c79d3c1d34bec8063"} Sep 29 20:42:58 crc kubenswrapper[4741]: I0929 20:42:58.259751 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7020456ac76fc76619766e2cf63e3e6f65fbc223f39f2e5c79d3c1d34bec8063" Sep 29 20:42:58 crc kubenswrapper[4741]: I0929 20:42:58.259848 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dc4f2" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.282415 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-81ca-account-create-jhlwq"] Sep 29 20:43:04 crc kubenswrapper[4741]: E0929 20:43:04.283671 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93df880-9269-48e4-9815-bbd513579722" containerName="mariadb-database-create" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.283696 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93df880-9269-48e4-9815-bbd513579722" containerName="mariadb-database-create" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.284052 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e93df880-9269-48e4-9815-bbd513579722" containerName="mariadb-database-create" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.285062 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-81ca-account-create-jhlwq" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.287125 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-81ca-account-create-jhlwq"] Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.287679 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.389299 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cvs5\" (UniqueName: \"kubernetes.io/projected/e4f3900c-8af5-484e-b0f3-3a60fdcc33e8-kube-api-access-7cvs5\") pod \"cinder-81ca-account-create-jhlwq\" (UID: \"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8\") " pod="openstack/cinder-81ca-account-create-jhlwq" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.491826 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cvs5\" (UniqueName: \"kubernetes.io/projected/e4f3900c-8af5-484e-b0f3-3a60fdcc33e8-kube-api-access-7cvs5\") pod \"cinder-81ca-account-create-jhlwq\" (UID: \"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8\") " pod="openstack/cinder-81ca-account-create-jhlwq" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.512010 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cvs5\" (UniqueName: \"kubernetes.io/projected/e4f3900c-8af5-484e-b0f3-3a60fdcc33e8-kube-api-access-7cvs5\") pod \"cinder-81ca-account-create-jhlwq\" (UID: \"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8\") " pod="openstack/cinder-81ca-account-create-jhlwq" Sep 29 20:43:04 crc kubenswrapper[4741]: I0929 20:43:04.610251 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-81ca-account-create-jhlwq" Sep 29 20:43:05 crc kubenswrapper[4741]: I0929 20:43:05.061759 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-81ca-account-create-jhlwq"] Sep 29 20:43:05 crc kubenswrapper[4741]: I0929 20:43:05.332141 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-81ca-account-create-jhlwq" event={"ID":"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8","Type":"ContainerStarted","Data":"27c279c151855c3f97e1ff2e9ad21c728949a8336c84380bee8b4e58ef7fad93"} Sep 29 20:43:05 crc kubenswrapper[4741]: I0929 20:43:05.332193 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-81ca-account-create-jhlwq" event={"ID":"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8","Type":"ContainerStarted","Data":"f136f327a3eaa2f5750c478ed02fee08674ba92606b0c5188af7d2f0381cc85c"} Sep 29 20:43:06 crc kubenswrapper[4741]: I0929 20:43:06.341546 4741 generic.go:334] "Generic (PLEG): container finished" podID="e4f3900c-8af5-484e-b0f3-3a60fdcc33e8" containerID="27c279c151855c3f97e1ff2e9ad21c728949a8336c84380bee8b4e58ef7fad93" exitCode=0 Sep 29 20:43:06 crc kubenswrapper[4741]: I0929 20:43:06.342148 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-81ca-account-create-jhlwq" event={"ID":"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8","Type":"ContainerDied","Data":"27c279c151855c3f97e1ff2e9ad21c728949a8336c84380bee8b4e58ef7fad93"} Sep 29 20:43:06 crc kubenswrapper[4741]: I0929 20:43:06.683244 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-81ca-account-create-jhlwq" Sep 29 20:43:06 crc kubenswrapper[4741]: I0929 20:43:06.749974 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cvs5\" (UniqueName: \"kubernetes.io/projected/e4f3900c-8af5-484e-b0f3-3a60fdcc33e8-kube-api-access-7cvs5\") pod \"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8\" (UID: \"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8\") " Sep 29 20:43:06 crc kubenswrapper[4741]: I0929 20:43:06.757111 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f3900c-8af5-484e-b0f3-3a60fdcc33e8-kube-api-access-7cvs5" (OuterVolumeSpecName: "kube-api-access-7cvs5") pod "e4f3900c-8af5-484e-b0f3-3a60fdcc33e8" (UID: "e4f3900c-8af5-484e-b0f3-3a60fdcc33e8"). InnerVolumeSpecName "kube-api-access-7cvs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:06 crc kubenswrapper[4741]: I0929 20:43:06.856145 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cvs5\" (UniqueName: \"kubernetes.io/projected/e4f3900c-8af5-484e-b0f3-3a60fdcc33e8-kube-api-access-7cvs5\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:07 crc kubenswrapper[4741]: I0929 20:43:07.354471 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-81ca-account-create-jhlwq" event={"ID":"e4f3900c-8af5-484e-b0f3-3a60fdcc33e8","Type":"ContainerDied","Data":"f136f327a3eaa2f5750c478ed02fee08674ba92606b0c5188af7d2f0381cc85c"} Sep 29 20:43:07 crc kubenswrapper[4741]: I0929 20:43:07.354518 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f136f327a3eaa2f5750c478ed02fee08674ba92606b0c5188af7d2f0381cc85c" Sep 29 20:43:07 crc kubenswrapper[4741]: I0929 20:43:07.354542 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-81ca-account-create-jhlwq" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.443877 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-d9twl"] Sep 29 20:43:09 crc kubenswrapper[4741]: E0929 20:43:09.445318 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f3900c-8af5-484e-b0f3-3a60fdcc33e8" containerName="mariadb-account-create" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.445353 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f3900c-8af5-484e-b0f3-3a60fdcc33e8" containerName="mariadb-account-create" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.445693 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f3900c-8af5-484e-b0f3-3a60fdcc33e8" containerName="mariadb-account-create" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.446286 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.448782 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.449315 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.450682 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-w92tc" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.472935 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-d9twl"] Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.604320 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-scripts\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.604372 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0dd2e44-2577-4579-911e-6a75c3063da0-etc-machine-id\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.604523 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dkr9\" (UniqueName: \"kubernetes.io/projected/a0dd2e44-2577-4579-911e-6a75c3063da0-kube-api-access-4dkr9\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.604596 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-db-sync-config-data\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.604639 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-config-data\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.604740 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-combined-ca-bundle\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.706217 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-scripts\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.706274 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0dd2e44-2577-4579-911e-6a75c3063da0-etc-machine-id\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.706339 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dkr9\" (UniqueName: \"kubernetes.io/projected/a0dd2e44-2577-4579-911e-6a75c3063da0-kube-api-access-4dkr9\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.706383 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-db-sync-config-data\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.706417 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-config-data\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.706444 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-combined-ca-bundle\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.706435 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0dd2e44-2577-4579-911e-6a75c3063da0-etc-machine-id\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.711940 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-config-data\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.713664 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-combined-ca-bundle\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.717718 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-scripts\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.717759 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-db-sync-config-data\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.722273 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dkr9\" (UniqueName: \"kubernetes.io/projected/a0dd2e44-2577-4579-911e-6a75c3063da0-kube-api-access-4dkr9\") pod \"cinder-db-sync-d9twl\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:09 crc kubenswrapper[4741]: I0929 20:43:09.775226 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:10 crc kubenswrapper[4741]: I0929 20:43:10.324892 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-d9twl"] Sep 29 20:43:10 crc kubenswrapper[4741]: I0929 20:43:10.392589 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d9twl" event={"ID":"a0dd2e44-2577-4579-911e-6a75c3063da0","Type":"ContainerStarted","Data":"1b75b95f3cd30cd355be4ff4f5fc8387c64bca07ef43ac29b38ec0e951e975ce"} Sep 29 20:43:11 crc kubenswrapper[4741]: I0929 20:43:11.401347 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d9twl" event={"ID":"a0dd2e44-2577-4579-911e-6a75c3063da0","Type":"ContainerStarted","Data":"0ec01ac2e2528dd738c8abcd16735488f6b79156508f049af4d769d31f1396b7"} Sep 29 20:43:11 crc kubenswrapper[4741]: I0929 20:43:11.417318 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-d9twl" podStartSLOduration=2.417296592 podStartE2EDuration="2.417296592s" podCreationTimestamp="2025-09-29 20:43:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:11.415246598 +0000 UTC m=+5633.063035940" watchObservedRunningTime="2025-09-29 20:43:11.417296592 +0000 UTC m=+5633.065085924" Sep 29 20:43:13 crc kubenswrapper[4741]: I0929 20:43:13.424083 4741 generic.go:334] "Generic (PLEG): container finished" podID="a0dd2e44-2577-4579-911e-6a75c3063da0" containerID="0ec01ac2e2528dd738c8abcd16735488f6b79156508f049af4d769d31f1396b7" exitCode=0 Sep 29 20:43:13 crc kubenswrapper[4741]: I0929 20:43:13.424155 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d9twl" event={"ID":"a0dd2e44-2577-4579-911e-6a75c3063da0","Type":"ContainerDied","Data":"0ec01ac2e2528dd738c8abcd16735488f6b79156508f049af4d769d31f1396b7"} Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.820490 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.908460 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-config-data\") pod \"a0dd2e44-2577-4579-911e-6a75c3063da0\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.908594 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dkr9\" (UniqueName: \"kubernetes.io/projected/a0dd2e44-2577-4579-911e-6a75c3063da0-kube-api-access-4dkr9\") pod \"a0dd2e44-2577-4579-911e-6a75c3063da0\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.908634 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-scripts\") pod \"a0dd2e44-2577-4579-911e-6a75c3063da0\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.908697 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-combined-ca-bundle\") pod \"a0dd2e44-2577-4579-911e-6a75c3063da0\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.908717 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0dd2e44-2577-4579-911e-6a75c3063da0-etc-machine-id\") pod \"a0dd2e44-2577-4579-911e-6a75c3063da0\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.908743 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-db-sync-config-data\") pod \"a0dd2e44-2577-4579-911e-6a75c3063da0\" (UID: \"a0dd2e44-2577-4579-911e-6a75c3063da0\") " Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.910490 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0dd2e44-2577-4579-911e-6a75c3063da0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a0dd2e44-2577-4579-911e-6a75c3063da0" (UID: "a0dd2e44-2577-4579-911e-6a75c3063da0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.917303 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-scripts" (OuterVolumeSpecName: "scripts") pod "a0dd2e44-2577-4579-911e-6a75c3063da0" (UID: "a0dd2e44-2577-4579-911e-6a75c3063da0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.917324 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a0dd2e44-2577-4579-911e-6a75c3063da0" (UID: "a0dd2e44-2577-4579-911e-6a75c3063da0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.922590 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0dd2e44-2577-4579-911e-6a75c3063da0-kube-api-access-4dkr9" (OuterVolumeSpecName: "kube-api-access-4dkr9") pod "a0dd2e44-2577-4579-911e-6a75c3063da0" (UID: "a0dd2e44-2577-4579-911e-6a75c3063da0"). InnerVolumeSpecName "kube-api-access-4dkr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.968712 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0dd2e44-2577-4579-911e-6a75c3063da0" (UID: "a0dd2e44-2577-4579-911e-6a75c3063da0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:14 crc kubenswrapper[4741]: I0929 20:43:14.980481 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-config-data" (OuterVolumeSpecName: "config-data") pod "a0dd2e44-2577-4579-911e-6a75c3063da0" (UID: "a0dd2e44-2577-4579-911e-6a75c3063da0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.011058 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dkr9\" (UniqueName: \"kubernetes.io/projected/a0dd2e44-2577-4579-911e-6a75c3063da0-kube-api-access-4dkr9\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.011088 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.011100 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.011109 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0dd2e44-2577-4579-911e-6a75c3063da0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.011118 4741 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.011126 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dd2e44-2577-4579-911e-6a75c3063da0-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.446386 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d9twl" event={"ID":"a0dd2e44-2577-4579-911e-6a75c3063da0","Type":"ContainerDied","Data":"1b75b95f3cd30cd355be4ff4f5fc8387c64bca07ef43ac29b38ec0e951e975ce"} Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.446948 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b75b95f3cd30cd355be4ff4f5fc8387c64bca07ef43ac29b38ec0e951e975ce" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.446455 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d9twl" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.789781 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-866bddffb7-6rh5n"] Sep 29 20:43:15 crc kubenswrapper[4741]: E0929 20:43:15.790380 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0dd2e44-2577-4579-911e-6a75c3063da0" containerName="cinder-db-sync" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.790403 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0dd2e44-2577-4579-911e-6a75c3063da0" containerName="cinder-db-sync" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.790610 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0dd2e44-2577-4579-911e-6a75c3063da0" containerName="cinder-db-sync" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.796938 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.838458 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-866bddffb7-6rh5n"] Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.925961 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxk7c\" (UniqueName: \"kubernetes.io/projected/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-kube-api-access-xxk7c\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.926064 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-config\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.926152 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-sb\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.926238 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-nb\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.926304 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-dns-svc\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.930004 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.932020 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.934868 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.934900 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.935151 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-w92tc" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.935344 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 29 20:43:15 crc kubenswrapper[4741]: I0929 20:43:15.943589 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.028101 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da521cf-10ef-423e-8649-6452fb96f1d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.028186 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-sb\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.028337 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-nb\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.028467 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-dns-svc\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.028624 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.028770 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-scripts\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.028954 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029216 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxk7c\" (UniqueName: \"kubernetes.io/projected/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-kube-api-access-xxk7c\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029335 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6da521cf-10ef-423e-8649-6452fb96f1d2-logs\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029421 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-dns-svc\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029449 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-config\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029474 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-sb\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029488 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029516 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9mlh\" (UniqueName: \"kubernetes.io/projected/6da521cf-10ef-423e-8649-6452fb96f1d2-kube-api-access-c9mlh\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.029718 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-nb\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.030296 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-config\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.055315 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxk7c\" (UniqueName: \"kubernetes.io/projected/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-kube-api-access-xxk7c\") pod \"dnsmasq-dns-866bddffb7-6rh5n\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.130849 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.130914 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-scripts\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.130960 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.130998 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6da521cf-10ef-423e-8649-6452fb96f1d2-logs\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.131022 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.131038 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9mlh\" (UniqueName: \"kubernetes.io/projected/6da521cf-10ef-423e-8649-6452fb96f1d2-kube-api-access-c9mlh\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.131062 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da521cf-10ef-423e-8649-6452fb96f1d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.131174 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da521cf-10ef-423e-8649-6452fb96f1d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.131952 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6da521cf-10ef-423e-8649-6452fb96f1d2-logs\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.136941 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.137234 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-scripts\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.137354 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.139126 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.147962 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9mlh\" (UniqueName: \"kubernetes.io/projected/6da521cf-10ef-423e-8649-6452fb96f1d2-kube-api-access-c9mlh\") pod \"cinder-api-0\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.194856 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.259320 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.694414 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-866bddffb7-6rh5n"] Sep 29 20:43:16 crc kubenswrapper[4741]: I0929 20:43:16.815040 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:16 crc kubenswrapper[4741]: W0929 20:43:16.818498 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6da521cf_10ef_423e_8649_6452fb96f1d2.slice/crio-f0f9d90d447776c3f298c647a65bebc73a432956b5e66f33bb47c2094794135b WatchSource:0}: Error finding container f0f9d90d447776c3f298c647a65bebc73a432956b5e66f33bb47c2094794135b: Status 404 returned error can't find the container with id f0f9d90d447776c3f298c647a65bebc73a432956b5e66f33bb47c2094794135b Sep 29 20:43:17 crc kubenswrapper[4741]: I0929 20:43:17.468989 4741 generic.go:334] "Generic (PLEG): container finished" podID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerID="6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe" exitCode=0 Sep 29 20:43:17 crc kubenswrapper[4741]: I0929 20:43:17.469038 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" event={"ID":"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3","Type":"ContainerDied","Data":"6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe"} Sep 29 20:43:17 crc kubenswrapper[4741]: I0929 20:43:17.469353 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" event={"ID":"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3","Type":"ContainerStarted","Data":"4bb816675ff295e7de37a28814d4c06fe477f2b6727c1f3e4ff1bdddada65124"} Sep 29 20:43:17 crc kubenswrapper[4741]: I0929 20:43:17.471360 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6da521cf-10ef-423e-8649-6452fb96f1d2","Type":"ContainerStarted","Data":"d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30"} Sep 29 20:43:17 crc kubenswrapper[4741]: I0929 20:43:17.471479 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6da521cf-10ef-423e-8649-6452fb96f1d2","Type":"ContainerStarted","Data":"f0f9d90d447776c3f298c647a65bebc73a432956b5e66f33bb47c2094794135b"} Sep 29 20:43:18 crc kubenswrapper[4741]: I0929 20:43:18.482318 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6da521cf-10ef-423e-8649-6452fb96f1d2","Type":"ContainerStarted","Data":"ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c"} Sep 29 20:43:18 crc kubenswrapper[4741]: I0929 20:43:18.482718 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 29 20:43:18 crc kubenswrapper[4741]: I0929 20:43:18.485578 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" event={"ID":"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3","Type":"ContainerStarted","Data":"a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a"} Sep 29 20:43:18 crc kubenswrapper[4741]: I0929 20:43:18.485762 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:18 crc kubenswrapper[4741]: I0929 20:43:18.504055 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.504035057 podStartE2EDuration="3.504035057s" podCreationTimestamp="2025-09-29 20:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:18.499969991 +0000 UTC m=+5640.147759343" watchObservedRunningTime="2025-09-29 20:43:18.504035057 +0000 UTC m=+5640.151824399" Sep 29 20:43:18 crc kubenswrapper[4741]: I0929 20:43:18.528595 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" podStartSLOduration=3.528572996 podStartE2EDuration="3.528572996s" podCreationTimestamp="2025-09-29 20:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:18.519058502 +0000 UTC m=+5640.166847844" watchObservedRunningTime="2025-09-29 20:43:18.528572996 +0000 UTC m=+5640.176362338" Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.196550 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.257747 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d8cc9c9c-8vpqg"] Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.258026 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" podUID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerName="dnsmasq-dns" containerID="cri-o://f9e835d64da2e05f43f3fd7a734d412f4da9ae5eac2a36ff86c9341bc1ad0c78" gracePeriod=10 Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.551814 4741 generic.go:334] "Generic (PLEG): container finished" podID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerID="f9e835d64da2e05f43f3fd7a734d412f4da9ae5eac2a36ff86c9341bc1ad0c78" exitCode=0 Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.551859 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" event={"ID":"d20b7eec-51aa-4fbf-a6ec-58f05a97e540","Type":"ContainerDied","Data":"f9e835d64da2e05f43f3fd7a734d412f4da9ae5eac2a36ff86c9341bc1ad0c78"} Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.833052 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.930832 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-nb\") pod \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.930986 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-config\") pod \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.931137 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-dns-svc\") pod \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.931166 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5w5b\" (UniqueName: \"kubernetes.io/projected/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-kube-api-access-l5w5b\") pod \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.931185 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-sb\") pod \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\" (UID: \"d20b7eec-51aa-4fbf-a6ec-58f05a97e540\") " Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.947503 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-kube-api-access-l5w5b" (OuterVolumeSpecName: "kube-api-access-l5w5b") pod "d20b7eec-51aa-4fbf-a6ec-58f05a97e540" (UID: "d20b7eec-51aa-4fbf-a6ec-58f05a97e540"). InnerVolumeSpecName "kube-api-access-l5w5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.983698 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d20b7eec-51aa-4fbf-a6ec-58f05a97e540" (UID: "d20b7eec-51aa-4fbf-a6ec-58f05a97e540"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.985027 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d20b7eec-51aa-4fbf-a6ec-58f05a97e540" (UID: "d20b7eec-51aa-4fbf-a6ec-58f05a97e540"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:43:26 crc kubenswrapper[4741]: I0929 20:43:26.993119 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-config" (OuterVolumeSpecName: "config") pod "d20b7eec-51aa-4fbf-a6ec-58f05a97e540" (UID: "d20b7eec-51aa-4fbf-a6ec-58f05a97e540"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.009284 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d20b7eec-51aa-4fbf-a6ec-58f05a97e540" (UID: "d20b7eec-51aa-4fbf-a6ec-58f05a97e540"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.033151 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.033188 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5w5b\" (UniqueName: \"kubernetes.io/projected/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-kube-api-access-l5w5b\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.033203 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.033216 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.033226 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20b7eec-51aa-4fbf-a6ec-58f05a97e540-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.561069 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" event={"ID":"d20b7eec-51aa-4fbf-a6ec-58f05a97e540","Type":"ContainerDied","Data":"f8e412083e0526090bd6dac805d4db360fb1a2e6019a86bd3f8ca7307552ab31"} Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.561115 4741 scope.go:117] "RemoveContainer" containerID="f9e835d64da2e05f43f3fd7a734d412f4da9ae5eac2a36ff86c9341bc1ad0c78" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.561217 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d8cc9c9c-8vpqg" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.582430 4741 scope.go:117] "RemoveContainer" containerID="b941aef73311cb3c9242cfc4a773bef4fc1a78d72397c2623c5b2398840b5989" Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.585371 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d8cc9c9c-8vpqg"] Sep 29 20:43:27 crc kubenswrapper[4741]: I0929 20:43:27.592615 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d8cc9c9c-8vpqg"] Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.140257 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.303337 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.303583 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d813c027-ccbb-41f2-8ae1-4e547b1ad080" containerName="nova-scheduler-scheduler" containerID="cri-o://8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.321300 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.321617 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="11c6a275-bb37-4f95-a582-2406799d11c1" containerName="nova-cell0-conductor-conductor" containerID="cri-o://bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.329440 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.329684 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="cb20babd-3a68-4c55-a6f6-2da696e4044a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.338579 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.338777 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-log" containerID="cri-o://e1e215ca44b835d24fe577aa39c784d4c217a95d0c263059037cb67e6daa7178" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.339149 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-api" containerID="cri-o://3d842fabb3d53f20f0e136863c5ebb114fb5e99f3814b59d00241ba9fe2fffba" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.345044 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.345250 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-log" containerID="cri-o://20bc3f9ead0984f7f0fb2ab8727a8909152dc8949fa6c4c13a884b281865cc78" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.345381 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-metadata" containerID="cri-o://96dc27c4a2c3bd098e5de6a697b6cd4762f6e2da63b55b540d548e98b85ecca5" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.405106 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.405352 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="c989f82b-2539-45be-8138-148762687ef0" containerName="nova-cell1-conductor-conductor" containerID="cri-o://a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" gracePeriod=30 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.569163 4741 generic.go:334] "Generic (PLEG): container finished" podID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerID="e1e215ca44b835d24fe577aa39c784d4c217a95d0c263059037cb67e6daa7178" exitCode=143 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.569242 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff279576-11ac-4c99-9b06-934b9c6e2257","Type":"ContainerDied","Data":"e1e215ca44b835d24fe577aa39c784d4c217a95d0c263059037cb67e6daa7178"} Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.571610 4741 generic.go:334] "Generic (PLEG): container finished" podID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerID="20bc3f9ead0984f7f0fb2ab8727a8909152dc8949fa6c4c13a884b281865cc78" exitCode=143 Sep 29 20:43:28 crc kubenswrapper[4741]: I0929 20:43:28.571697 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"30f20bc6-75ac-42bf-957b-37b925499cd2","Type":"ContainerDied","Data":"20bc3f9ead0984f7f0fb2ab8727a8909152dc8949fa6c4c13a884b281865cc78"} Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.054199 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.078074 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmjfg\" (UniqueName: \"kubernetes.io/projected/cb20babd-3a68-4c55-a6f6-2da696e4044a-kube-api-access-fmjfg\") pod \"cb20babd-3a68-4c55-a6f6-2da696e4044a\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.078120 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-config-data\") pod \"cb20babd-3a68-4c55-a6f6-2da696e4044a\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.078174 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-combined-ca-bundle\") pod \"cb20babd-3a68-4c55-a6f6-2da696e4044a\" (UID: \"cb20babd-3a68-4c55-a6f6-2da696e4044a\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.086583 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb20babd-3a68-4c55-a6f6-2da696e4044a-kube-api-access-fmjfg" (OuterVolumeSpecName: "kube-api-access-fmjfg") pod "cb20babd-3a68-4c55-a6f6-2da696e4044a" (UID: "cb20babd-3a68-4c55-a6f6-2da696e4044a"). InnerVolumeSpecName "kube-api-access-fmjfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.103680 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" path="/var/lib/kubelet/pods/d20b7eec-51aa-4fbf-a6ec-58f05a97e540/volumes" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.109842 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb20babd-3a68-4c55-a6f6-2da696e4044a" (UID: "cb20babd-3a68-4c55-a6f6-2da696e4044a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.112244 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-config-data" (OuterVolumeSpecName: "config-data") pod "cb20babd-3a68-4c55-a6f6-2da696e4044a" (UID: "cb20babd-3a68-4c55-a6f6-2da696e4044a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.185762 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmjfg\" (UniqueName: \"kubernetes.io/projected/cb20babd-3a68-4c55-a6f6-2da696e4044a-kube-api-access-fmjfg\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.185791 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.185802 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb20babd-3a68-4c55-a6f6-2da696e4044a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.251627 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.287020 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7bnw\" (UniqueName: \"kubernetes.io/projected/d813c027-ccbb-41f2-8ae1-4e547b1ad080-kube-api-access-l7bnw\") pod \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.287061 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-combined-ca-bundle\") pod \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.287224 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-config-data\") pod \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\" (UID: \"d813c027-ccbb-41f2-8ae1-4e547b1ad080\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.292245 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d813c027-ccbb-41f2-8ae1-4e547b1ad080-kube-api-access-l7bnw" (OuterVolumeSpecName: "kube-api-access-l7bnw") pod "d813c027-ccbb-41f2-8ae1-4e547b1ad080" (UID: "d813c027-ccbb-41f2-8ae1-4e547b1ad080"). InnerVolumeSpecName "kube-api-access-l7bnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.315093 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-config-data" (OuterVolumeSpecName: "config-data") pod "d813c027-ccbb-41f2-8ae1-4e547b1ad080" (UID: "d813c027-ccbb-41f2-8ae1-4e547b1ad080"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.315487 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.319549 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d813c027-ccbb-41f2-8ae1-4e547b1ad080" (UID: "d813c027-ccbb-41f2-8ae1-4e547b1ad080"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.388548 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6dtq\" (UniqueName: \"kubernetes.io/projected/11c6a275-bb37-4f95-a582-2406799d11c1-kube-api-access-x6dtq\") pod \"11c6a275-bb37-4f95-a582-2406799d11c1\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.388631 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-combined-ca-bundle\") pod \"11c6a275-bb37-4f95-a582-2406799d11c1\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.388760 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-config-data\") pod \"11c6a275-bb37-4f95-a582-2406799d11c1\" (UID: \"11c6a275-bb37-4f95-a582-2406799d11c1\") " Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.389144 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.389161 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7bnw\" (UniqueName: \"kubernetes.io/projected/d813c027-ccbb-41f2-8ae1-4e547b1ad080-kube-api-access-l7bnw\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.389170 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d813c027-ccbb-41f2-8ae1-4e547b1ad080-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.391655 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11c6a275-bb37-4f95-a582-2406799d11c1-kube-api-access-x6dtq" (OuterVolumeSpecName: "kube-api-access-x6dtq") pod "11c6a275-bb37-4f95-a582-2406799d11c1" (UID: "11c6a275-bb37-4f95-a582-2406799d11c1"). InnerVolumeSpecName "kube-api-access-x6dtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.412618 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-config-data" (OuterVolumeSpecName: "config-data") pod "11c6a275-bb37-4f95-a582-2406799d11c1" (UID: "11c6a275-bb37-4f95-a582-2406799d11c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.414101 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11c6a275-bb37-4f95-a582-2406799d11c1" (UID: "11c6a275-bb37-4f95-a582-2406799d11c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.491175 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6dtq\" (UniqueName: \"kubernetes.io/projected/11c6a275-bb37-4f95-a582-2406799d11c1-kube-api-access-x6dtq\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.491293 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.491369 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c6a275-bb37-4f95-a582-2406799d11c1-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.581248 4741 generic.go:334] "Generic (PLEG): container finished" podID="cb20babd-3a68-4c55-a6f6-2da696e4044a" containerID="6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1" exitCode=0 Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.581345 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.581409 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb20babd-3a68-4c55-a6f6-2da696e4044a","Type":"ContainerDied","Data":"6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1"} Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.582308 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb20babd-3a68-4c55-a6f6-2da696e4044a","Type":"ContainerDied","Data":"8a39538de64cda6d6be93145893702e399ae3193ab518c9c65cb3d90597ccc79"} Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.582340 4741 scope.go:117] "RemoveContainer" containerID="6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.584380 4741 generic.go:334] "Generic (PLEG): container finished" podID="d813c027-ccbb-41f2-8ae1-4e547b1ad080" containerID="8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d" exitCode=0 Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.584472 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d813c027-ccbb-41f2-8ae1-4e547b1ad080","Type":"ContainerDied","Data":"8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d"} Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.584497 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d813c027-ccbb-41f2-8ae1-4e547b1ad080","Type":"ContainerDied","Data":"9ad0edd31fcefc4c79019b0b6c8175a12217bdb1ceaed1843224e5259160ac96"} Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.585777 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.587155 4741 generic.go:334] "Generic (PLEG): container finished" podID="11c6a275-bb37-4f95-a582-2406799d11c1" containerID="bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c" exitCode=0 Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.587189 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"11c6a275-bb37-4f95-a582-2406799d11c1","Type":"ContainerDied","Data":"bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c"} Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.587216 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"11c6a275-bb37-4f95-a582-2406799d11c1","Type":"ContainerDied","Data":"0deb5ab5ecfcae38c9a5fa92818d47187501ba22efacf716d388f200bae48fb6"} Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.587259 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.610277 4741 scope.go:117] "RemoveContainer" containerID="6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1" Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.610871 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1\": container with ID starting with 6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1 not found: ID does not exist" containerID="6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.610908 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1"} err="failed to get container status \"6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1\": rpc error: code = NotFound desc = could not find container \"6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1\": container with ID starting with 6991eff9f2eddf231df91c07495a4a8f7f612714aee104dc867f53c2b683c8b1 not found: ID does not exist" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.610939 4741 scope.go:117] "RemoveContainer" containerID="8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.633343 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.644567 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.656762 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.673584 4741 scope.go:117] "RemoveContainer" containerID="8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d" Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.674733 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d\": container with ID starting with 8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d not found: ID does not exist" containerID="8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.674764 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d"} err="failed to get container status \"8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d\": rpc error: code = NotFound desc = could not find container \"8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d\": container with ID starting with 8bc83ade365613e24d0b2fbbf723aa64c4c976f25d943cec0032b8374e24a23d not found: ID does not exist" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.674782 4741 scope.go:117] "RemoveContainer" containerID="bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.700300 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.713437 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.713894 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb20babd-3a68-4c55-a6f6-2da696e4044a" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.713910 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb20babd-3a68-4c55-a6f6-2da696e4044a" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.713934 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerName="init" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.713942 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerName="init" Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.713965 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d813c027-ccbb-41f2-8ae1-4e547b1ad080" containerName="nova-scheduler-scheduler" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.713975 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d813c027-ccbb-41f2-8ae1-4e547b1ad080" containerName="nova-scheduler-scheduler" Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.713986 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c6a275-bb37-4f95-a582-2406799d11c1" containerName="nova-cell0-conductor-conductor" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.713993 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c6a275-bb37-4f95-a582-2406799d11c1" containerName="nova-cell0-conductor-conductor" Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.714012 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerName="dnsmasq-dns" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.714019 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerName="dnsmasq-dns" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.714236 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb20babd-3a68-4c55-a6f6-2da696e4044a" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.714250 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d813c027-ccbb-41f2-8ae1-4e547b1ad080" containerName="nova-scheduler-scheduler" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.714288 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20b7eec-51aa-4fbf-a6ec-58f05a97e540" containerName="dnsmasq-dns" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.714303 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c6a275-bb37-4f95-a582-2406799d11c1" containerName="nova-cell0-conductor-conductor" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.717062 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.717283 4741 scope.go:117] "RemoveContainer" containerID="bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c" Sep 29 20:43:29 crc kubenswrapper[4741]: E0929 20:43:29.718240 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c\": container with ID starting with bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c not found: ID does not exist" containerID="bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.718307 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c"} err="failed to get container status \"bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c\": rpc error: code = NotFound desc = could not find container \"bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c\": container with ID starting with bdcef6d3276d1866e6eeb6a1261685c894617cded39f0302b2e7f4e1d232e83c not found: ID does not exist" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.720105 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.737781 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.758470 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.775465 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.791783 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.793015 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.795508 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.796584 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs22p\" (UniqueName: \"kubernetes.io/projected/3454c3e1-18f9-425f-a9fa-7403c89a128b-kube-api-access-zs22p\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.796623 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3454c3e1-18f9-425f-a9fa-7403c89a128b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.796668 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3454c3e1-18f9-425f-a9fa-7403c89a128b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.798710 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.811337 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.812593 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.814642 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.831016 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898548 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898612 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhbgk\" (UniqueName: \"kubernetes.io/projected/e3e16528-0ced-4c87-833b-9a09283689b5-kube-api-access-vhbgk\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898663 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898701 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-config-data\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898731 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs22p\" (UniqueName: \"kubernetes.io/projected/3454c3e1-18f9-425f-a9fa-7403c89a128b-kube-api-access-zs22p\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898761 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898787 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3454c3e1-18f9-425f-a9fa-7403c89a128b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898817 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3454c3e1-18f9-425f-a9fa-7403c89a128b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.898845 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zd7s\" (UniqueName: \"kubernetes.io/projected/d8026026-9009-4a11-b671-d6d071a6ad7e-kube-api-access-8zd7s\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.903612 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3454c3e1-18f9-425f-a9fa-7403c89a128b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.905920 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3454c3e1-18f9-425f-a9fa-7403c89a128b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:29 crc kubenswrapper[4741]: I0929 20:43:29.928407 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs22p\" (UniqueName: \"kubernetes.io/projected/3454c3e1-18f9-425f-a9fa-7403c89a128b-kube-api-access-zs22p\") pod \"nova-cell1-novncproxy-0\" (UID: \"3454c3e1-18f9-425f-a9fa-7403c89a128b\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.001072 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.001575 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zd7s\" (UniqueName: \"kubernetes.io/projected/d8026026-9009-4a11-b671-d6d071a6ad7e-kube-api-access-8zd7s\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.001743 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.001823 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhbgk\" (UniqueName: \"kubernetes.io/projected/e3e16528-0ced-4c87-833b-9a09283689b5-kube-api-access-vhbgk\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.001907 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.001988 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-config-data\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.006333 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-config-data\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.007033 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.008694 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.016520 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.021773 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhbgk\" (UniqueName: \"kubernetes.io/projected/e3e16528-0ced-4c87-833b-9a09283689b5-kube-api-access-vhbgk\") pod \"nova-cell0-conductor-0\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.022860 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zd7s\" (UniqueName: \"kubernetes.io/projected/d8026026-9009-4a11-b671-d6d071a6ad7e-kube-api-access-8zd7s\") pod \"nova-scheduler-0\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " pod="openstack/nova-scheduler-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.053748 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.111002 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.133081 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.490761 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.598060 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3454c3e1-18f9-425f-a9fa-7403c89a128b","Type":"ContainerStarted","Data":"dddc958d886f701fda1dbd1749e919a3a71fa85ff3c39da1422f2b4248904436"} Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.641165 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 20:43:30 crc kubenswrapper[4741]: W0929 20:43:30.645800 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8026026_9009_4a11_b671_d6d071a6ad7e.slice/crio-2d10164788c3b18d922d1f9725776e3d7e60fe3b29d9c177a49e361c71125fb4 WatchSource:0}: Error finding container 2d10164788c3b18d922d1f9725776e3d7e60fe3b29d9c177a49e361c71125fb4: Status 404 returned error can't find the container with id 2d10164788c3b18d922d1f9725776e3d7e60fe3b29d9c177a49e361c71125fb4 Sep 29 20:43:30 crc kubenswrapper[4741]: I0929 20:43:30.647574 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.095752 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11c6a275-bb37-4f95-a582-2406799d11c1" path="/var/lib/kubelet/pods/11c6a275-bb37-4f95-a582-2406799d11c1/volumes" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.096643 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb20babd-3a68-4c55-a6f6-2da696e4044a" path="/var/lib/kubelet/pods/cb20babd-3a68-4c55-a6f6-2da696e4044a/volumes" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.097097 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d813c027-ccbb-41f2-8ae1-4e547b1ad080" path="/var/lib/kubelet/pods/d813c027-ccbb-41f2-8ae1-4e547b1ad080/volumes" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.493855 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": read tcp 10.217.0.2:57880->10.217.1.68:8775: read: connection reset by peer" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.494048 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": read tcp 10.217.0.2:57878->10.217.1.68:8775: read: connection reset by peer" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.633146 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d8026026-9009-4a11-b671-d6d071a6ad7e","Type":"ContainerStarted","Data":"a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23"} Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.633191 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d8026026-9009-4a11-b671-d6d071a6ad7e","Type":"ContainerStarted","Data":"2d10164788c3b18d922d1f9725776e3d7e60fe3b29d9c177a49e361c71125fb4"} Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.646870 4741 generic.go:334] "Generic (PLEG): container finished" podID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerID="3d842fabb3d53f20f0e136863c5ebb114fb5e99f3814b59d00241ba9fe2fffba" exitCode=0 Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.646929 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff279576-11ac-4c99-9b06-934b9c6e2257","Type":"ContainerDied","Data":"3d842fabb3d53f20f0e136863c5ebb114fb5e99f3814b59d00241ba9fe2fffba"} Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.648927 4741 generic.go:334] "Generic (PLEG): container finished" podID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerID="96dc27c4a2c3bd098e5de6a697b6cd4762f6e2da63b55b540d548e98b85ecca5" exitCode=0 Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.648965 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"30f20bc6-75ac-42bf-957b-37b925499cd2","Type":"ContainerDied","Data":"96dc27c4a2c3bd098e5de6a697b6cd4762f6e2da63b55b540d548e98b85ecca5"} Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.650015 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3454c3e1-18f9-425f-a9fa-7403c89a128b","Type":"ContainerStarted","Data":"b23325e171f8de934174a3e958b7a40e8aabfb8ed69700d01ddaa16826964e2c"} Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.663999 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e3e16528-0ced-4c87-833b-9a09283689b5","Type":"ContainerStarted","Data":"d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441"} Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.664031 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e3e16528-0ced-4c87-833b-9a09283689b5","Type":"ContainerStarted","Data":"c16ad00e0d4c8ab268a47839d5d0a3239bcb6612c3438221b0f0ff3e7a4ece58"} Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.664863 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.682868 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.682848329 podStartE2EDuration="2.682848329s" podCreationTimestamp="2025-09-29 20:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:31.657935037 +0000 UTC m=+5653.305724379" watchObservedRunningTime="2025-09-29 20:43:31.682848329 +0000 UTC m=+5653.330637661" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.701476 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.701451704 podStartE2EDuration="2.701451704s" podCreationTimestamp="2025-09-29 20:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:31.697500552 +0000 UTC m=+5653.345289884" watchObservedRunningTime="2025-09-29 20:43:31.701451704 +0000 UTC m=+5653.349241036" Sep 29 20:43:31 crc kubenswrapper[4741]: I0929 20:43:31.710828 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.710805724 podStartE2EDuration="2.710805724s" podCreationTimestamp="2025-09-29 20:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:31.684372786 +0000 UTC m=+5653.332162118" watchObservedRunningTime="2025-09-29 20:43:31.710805724 +0000 UTC m=+5653.358595056" Sep 29 20:43:31 crc kubenswrapper[4741]: E0929 20:43:31.984867 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 20:43:31 crc kubenswrapper[4741]: E0929 20:43:31.987810 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 20:43:32 crc kubenswrapper[4741]: E0929 20:43:31.996204 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Sep 29 20:43:32 crc kubenswrapper[4741]: E0929 20:43:31.996250 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="c989f82b-2539-45be-8138-148762687ef0" containerName="nova-cell1-conductor-conductor" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.036362 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.111693 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135018 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-config-data\") pod \"30f20bc6-75ac-42bf-957b-37b925499cd2\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135056 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpzh8\" (UniqueName: \"kubernetes.io/projected/ff279576-11ac-4c99-9b06-934b9c6e2257-kube-api-access-zpzh8\") pod \"ff279576-11ac-4c99-9b06-934b9c6e2257\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135082 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f20bc6-75ac-42bf-957b-37b925499cd2-logs\") pod \"30f20bc6-75ac-42bf-957b-37b925499cd2\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135100 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-config-data\") pod \"ff279576-11ac-4c99-9b06-934b9c6e2257\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135177 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd6pw\" (UniqueName: \"kubernetes.io/projected/30f20bc6-75ac-42bf-957b-37b925499cd2-kube-api-access-gd6pw\") pod \"30f20bc6-75ac-42bf-957b-37b925499cd2\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135199 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-combined-ca-bundle\") pod \"ff279576-11ac-4c99-9b06-934b9c6e2257\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135246 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff279576-11ac-4c99-9b06-934b9c6e2257-logs\") pod \"ff279576-11ac-4c99-9b06-934b9c6e2257\" (UID: \"ff279576-11ac-4c99-9b06-934b9c6e2257\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.135318 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-combined-ca-bundle\") pod \"30f20bc6-75ac-42bf-957b-37b925499cd2\" (UID: \"30f20bc6-75ac-42bf-957b-37b925499cd2\") " Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.136736 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff279576-11ac-4c99-9b06-934b9c6e2257-logs" (OuterVolumeSpecName: "logs") pod "ff279576-11ac-4c99-9b06-934b9c6e2257" (UID: "ff279576-11ac-4c99-9b06-934b9c6e2257"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.137108 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30f20bc6-75ac-42bf-957b-37b925499cd2-logs" (OuterVolumeSpecName: "logs") pod "30f20bc6-75ac-42bf-957b-37b925499cd2" (UID: "30f20bc6-75ac-42bf-957b-37b925499cd2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.144920 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30f20bc6-75ac-42bf-957b-37b925499cd2-kube-api-access-gd6pw" (OuterVolumeSpecName: "kube-api-access-gd6pw") pod "30f20bc6-75ac-42bf-957b-37b925499cd2" (UID: "30f20bc6-75ac-42bf-957b-37b925499cd2"). InnerVolumeSpecName "kube-api-access-gd6pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.146629 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff279576-11ac-4c99-9b06-934b9c6e2257-kube-api-access-zpzh8" (OuterVolumeSpecName: "kube-api-access-zpzh8") pod "ff279576-11ac-4c99-9b06-934b9c6e2257" (UID: "ff279576-11ac-4c99-9b06-934b9c6e2257"). InnerVolumeSpecName "kube-api-access-zpzh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.176697 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-config-data" (OuterVolumeSpecName: "config-data") pod "30f20bc6-75ac-42bf-957b-37b925499cd2" (UID: "30f20bc6-75ac-42bf-957b-37b925499cd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.184509 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-config-data" (OuterVolumeSpecName: "config-data") pod "ff279576-11ac-4c99-9b06-934b9c6e2257" (UID: "ff279576-11ac-4c99-9b06-934b9c6e2257"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.193218 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff279576-11ac-4c99-9b06-934b9c6e2257" (UID: "ff279576-11ac-4c99-9b06-934b9c6e2257"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.211879 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30f20bc6-75ac-42bf-957b-37b925499cd2" (UID: "30f20bc6-75ac-42bf-957b-37b925499cd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.237345 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd6pw\" (UniqueName: \"kubernetes.io/projected/30f20bc6-75ac-42bf-957b-37b925499cd2-kube-api-access-gd6pw\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.237531 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.237633 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff279576-11ac-4c99-9b06-934b9c6e2257-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.237712 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.237785 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f20bc6-75ac-42bf-957b-37b925499cd2-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.237843 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpzh8\" (UniqueName: \"kubernetes.io/projected/ff279576-11ac-4c99-9b06-934b9c6e2257-kube-api-access-zpzh8\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.237936 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f20bc6-75ac-42bf-957b-37b925499cd2-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.238047 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff279576-11ac-4c99-9b06-934b9c6e2257-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.690752 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.692404 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff279576-11ac-4c99-9b06-934b9c6e2257","Type":"ContainerDied","Data":"56a31fc65c75aa652940183723d69dc40ad61e6279d052251f3aef0d7a6dfc0b"} Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.692461 4741 scope.go:117] "RemoveContainer" containerID="3d842fabb3d53f20f0e136863c5ebb114fb5e99f3814b59d00241ba9fe2fffba" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.696510 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"30f20bc6-75ac-42bf-957b-37b925499cd2","Type":"ContainerDied","Data":"88a40d8642f6f3e121141148e11b223d5e5a41c0d235d3d4b2bf0beff687c2c6"} Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.696684 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.726203 4741 scope.go:117] "RemoveContainer" containerID="e1e215ca44b835d24fe577aa39c784d4c217a95d0c263059037cb67e6daa7178" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.752937 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.755725 4741 scope.go:117] "RemoveContainer" containerID="96dc27c4a2c3bd098e5de6a697b6cd4762f6e2da63b55b540d548e98b85ecca5" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.770685 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.784484 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.811915 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.818264 4741 scope.go:117] "RemoveContainer" containerID="20bc3f9ead0984f7f0fb2ab8727a8909152dc8949fa6c4c13a884b281865cc78" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.831370 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: E0929 20:43:32.831779 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-api" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.831796 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-api" Sep 29 20:43:32 crc kubenswrapper[4741]: E0929 20:43:32.831805 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-log" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.831813 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-log" Sep 29 20:43:32 crc kubenswrapper[4741]: E0929 20:43:32.831824 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-log" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.831830 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-log" Sep 29 20:43:32 crc kubenswrapper[4741]: E0929 20:43:32.831845 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-metadata" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.831851 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-metadata" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.832030 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-metadata" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.832044 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" containerName="nova-metadata-log" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.832055 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-api" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.832065 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" containerName="nova-api-log" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.833065 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.836499 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.839341 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.846357 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.847285 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-logs\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.847517 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-config-data\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.847548 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.847658 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slggh\" (UniqueName: \"kubernetes.io/projected/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-kube-api-access-slggh\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.848196 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.852541 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.859129 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.953978 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f394e16-a1d3-45a7-af53-53898e178300-logs\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954054 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-config-data\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954205 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slggh\" (UniqueName: \"kubernetes.io/projected/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-kube-api-access-slggh\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954272 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-logs\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954548 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954596 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-config-data\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954637 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljwp\" (UniqueName: \"kubernetes.io/projected/6f394e16-a1d3-45a7-af53-53898e178300-kube-api-access-8ljwp\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954656 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.954908 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-logs\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.970406 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.971205 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-config-data\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:32 crc kubenswrapper[4741]: I0929 20:43:32.971704 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slggh\" (UniqueName: \"kubernetes.io/projected/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-kube-api-access-slggh\") pod \"nova-api-0\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " pod="openstack/nova-api-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.055769 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f394e16-a1d3-45a7-af53-53898e178300-logs\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.055826 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-config-data\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.055922 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.055964 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljwp\" (UniqueName: \"kubernetes.io/projected/6f394e16-a1d3-45a7-af53-53898e178300-kube-api-access-8ljwp\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.056165 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f394e16-a1d3-45a7-af53-53898e178300-logs\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.059255 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-config-data\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.061019 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.072212 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljwp\" (UniqueName: \"kubernetes.io/projected/6f394e16-a1d3-45a7-af53-53898e178300-kube-api-access-8ljwp\") pod \"nova-metadata-0\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.101308 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30f20bc6-75ac-42bf-957b-37b925499cd2" path="/var/lib/kubelet/pods/30f20bc6-75ac-42bf-957b-37b925499cd2/volumes" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.101912 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff279576-11ac-4c99-9b06-934b9c6e2257" path="/var/lib/kubelet/pods/ff279576-11ac-4c99-9b06-934b9c6e2257/volumes" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.154755 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.166814 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.598626 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.680552 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 20:43:33 crc kubenswrapper[4741]: W0929 20:43:33.700672 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f394e16_a1d3_45a7_af53_53898e178300.slice/crio-93086df0952d7c43458a361aad0842c2bcf513681c8efa5677cbfa2993103c39 WatchSource:0}: Error finding container 93086df0952d7c43458a361aad0842c2bcf513681c8efa5677cbfa2993103c39: Status 404 returned error can't find the container with id 93086df0952d7c43458a361aad0842c2bcf513681c8efa5677cbfa2993103c39 Sep 29 20:43:33 crc kubenswrapper[4741]: I0929 20:43:33.724315 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ddafdfc-baf4-4368-b8c8-c3f5eafce192","Type":"ContainerStarted","Data":"8008bb5d40ddcec41d8036c81ccfe5e19df06fc4c7c90d615ce700d7503f5696"} Sep 29 20:43:34 crc kubenswrapper[4741]: I0929 20:43:34.736147 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f394e16-a1d3-45a7-af53-53898e178300","Type":"ContainerStarted","Data":"f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0"} Sep 29 20:43:34 crc kubenswrapper[4741]: I0929 20:43:34.736728 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f394e16-a1d3-45a7-af53-53898e178300","Type":"ContainerStarted","Data":"f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa"} Sep 29 20:43:34 crc kubenswrapper[4741]: I0929 20:43:34.736743 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f394e16-a1d3-45a7-af53-53898e178300","Type":"ContainerStarted","Data":"93086df0952d7c43458a361aad0842c2bcf513681c8efa5677cbfa2993103c39"} Sep 29 20:43:34 crc kubenswrapper[4741]: I0929 20:43:34.738737 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ddafdfc-baf4-4368-b8c8-c3f5eafce192","Type":"ContainerStarted","Data":"5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404"} Sep 29 20:43:34 crc kubenswrapper[4741]: I0929 20:43:34.738779 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ddafdfc-baf4-4368-b8c8-c3f5eafce192","Type":"ContainerStarted","Data":"6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441"} Sep 29 20:43:34 crc kubenswrapper[4741]: I0929 20:43:34.780206 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.780177074 podStartE2EDuration="2.780177074s" podCreationTimestamp="2025-09-29 20:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:34.757992127 +0000 UTC m=+5656.405781459" watchObservedRunningTime="2025-09-29 20:43:34.780177074 +0000 UTC m=+5656.427966416" Sep 29 20:43:34 crc kubenswrapper[4741]: I0929 20:43:34.783101 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.7830852740000003 podStartE2EDuration="2.783085274s" podCreationTimestamp="2025-09-29 20:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:34.779163153 +0000 UTC m=+5656.426952505" watchObservedRunningTime="2025-09-29 20:43:34.783085274 +0000 UTC m=+5656.430874606" Sep 29 20:43:35 crc kubenswrapper[4741]: I0929 20:43:35.055069 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:35 crc kubenswrapper[4741]: I0929 20:43:35.115257 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 20:43:35 crc kubenswrapper[4741]: I0929 20:43:35.163710 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.256113 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.326226 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5drj2\" (UniqueName: \"kubernetes.io/projected/c989f82b-2539-45be-8138-148762687ef0-kube-api-access-5drj2\") pod \"c989f82b-2539-45be-8138-148762687ef0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.326510 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-combined-ca-bundle\") pod \"c989f82b-2539-45be-8138-148762687ef0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.326637 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-config-data\") pod \"c989f82b-2539-45be-8138-148762687ef0\" (UID: \"c989f82b-2539-45be-8138-148762687ef0\") " Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.335578 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c989f82b-2539-45be-8138-148762687ef0-kube-api-access-5drj2" (OuterVolumeSpecName: "kube-api-access-5drj2") pod "c989f82b-2539-45be-8138-148762687ef0" (UID: "c989f82b-2539-45be-8138-148762687ef0"). InnerVolumeSpecName "kube-api-access-5drj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.360749 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c989f82b-2539-45be-8138-148762687ef0" (UID: "c989f82b-2539-45be-8138-148762687ef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.367008 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-config-data" (OuterVolumeSpecName: "config-data") pod "c989f82b-2539-45be-8138-148762687ef0" (UID: "c989f82b-2539-45be-8138-148762687ef0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.428912 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5drj2\" (UniqueName: \"kubernetes.io/projected/c989f82b-2539-45be-8138-148762687ef0-kube-api-access-5drj2\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.428957 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.428972 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c989f82b-2539-45be-8138-148762687ef0-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.758762 4741 generic.go:334] "Generic (PLEG): container finished" podID="c989f82b-2539-45be-8138-148762687ef0" containerID="a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" exitCode=0 Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.758803 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c989f82b-2539-45be-8138-148762687ef0","Type":"ContainerDied","Data":"a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d"} Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.758834 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c989f82b-2539-45be-8138-148762687ef0","Type":"ContainerDied","Data":"a90ed34654a93ddd7e71380b6873acb21bce5be97d2e65ec3d5238e611fd6697"} Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.758851 4741 scope.go:117] "RemoveContainer" containerID="a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.758874 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.781605 4741 scope.go:117] "RemoveContainer" containerID="a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" Sep 29 20:43:36 crc kubenswrapper[4741]: E0929 20:43:36.782982 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d\": container with ID starting with a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d not found: ID does not exist" containerID="a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.783027 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d"} err="failed to get container status \"a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d\": rpc error: code = NotFound desc = could not find container \"a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d\": container with ID starting with a6341ed4ee696aa6fdec195fdbb58f392593a6e292ed10a8f8a249e15c661a8d not found: ID does not exist" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.806773 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.818594 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.831230 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:43:36 crc kubenswrapper[4741]: E0929 20:43:36.831937 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c989f82b-2539-45be-8138-148762687ef0" containerName="nova-cell1-conductor-conductor" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.831958 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c989f82b-2539-45be-8138-148762687ef0" containerName="nova-cell1-conductor-conductor" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.832297 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c989f82b-2539-45be-8138-148762687ef0" containerName="nova-cell1-conductor-conductor" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.833318 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.837177 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.841705 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.938070 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.938127 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:36 crc kubenswrapper[4741]: I0929 20:43:36.938149 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgsjn\" (UniqueName: \"kubernetes.io/projected/546d78c4-996b-4f8d-af92-97982ba58890-kube-api-access-kgsjn\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.040054 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.040130 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.040151 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgsjn\" (UniqueName: \"kubernetes.io/projected/546d78c4-996b-4f8d-af92-97982ba58890-kube-api-access-kgsjn\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.044925 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.045676 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.060162 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgsjn\" (UniqueName: \"kubernetes.io/projected/546d78c4-996b-4f8d-af92-97982ba58890-kube-api-access-kgsjn\") pod \"nova-cell1-conductor-0\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.100527 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c989f82b-2539-45be-8138-148762687ef0" path="/var/lib/kubelet/pods/c989f82b-2539-45be-8138-148762687ef0/volumes" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.155751 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.585772 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 20:43:37 crc kubenswrapper[4741]: W0929 20:43:37.590491 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod546d78c4_996b_4f8d_af92_97982ba58890.slice/crio-0f272ec29b2560ae86ae8b52b7cd148e705a34ddc86bea17b5662248beaa9b07 WatchSource:0}: Error finding container 0f272ec29b2560ae86ae8b52b7cd148e705a34ddc86bea17b5662248beaa9b07: Status 404 returned error can't find the container with id 0f272ec29b2560ae86ae8b52b7cd148e705a34ddc86bea17b5662248beaa9b07 Sep 29 20:43:37 crc kubenswrapper[4741]: I0929 20:43:37.769222 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"546d78c4-996b-4f8d-af92-97982ba58890","Type":"ContainerStarted","Data":"0f272ec29b2560ae86ae8b52b7cd148e705a34ddc86bea17b5662248beaa9b07"} Sep 29 20:43:38 crc kubenswrapper[4741]: I0929 20:43:38.166993 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:43:38 crc kubenswrapper[4741]: I0929 20:43:38.167230 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 20:43:38 crc kubenswrapper[4741]: I0929 20:43:38.779255 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"546d78c4-996b-4f8d-af92-97982ba58890","Type":"ContainerStarted","Data":"88a52c252f6d9c79b2fc3dde07e626889fe77298cf38a87ab79861db10a63260"} Sep 29 20:43:38 crc kubenswrapper[4741]: I0929 20:43:38.779448 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:40 crc kubenswrapper[4741]: I0929 20:43:40.054599 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:40 crc kubenswrapper[4741]: I0929 20:43:40.064871 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:40 crc kubenswrapper[4741]: I0929 20:43:40.092968 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=4.092948267 podStartE2EDuration="4.092948267s" podCreationTimestamp="2025-09-29 20:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:38.802714159 +0000 UTC m=+5660.450503491" watchObservedRunningTime="2025-09-29 20:43:40.092948267 +0000 UTC m=+5661.740737609" Sep 29 20:43:40 crc kubenswrapper[4741]: I0929 20:43:40.111959 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 20:43:40 crc kubenswrapper[4741]: I0929 20:43:40.153903 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 20:43:40 crc kubenswrapper[4741]: I0929 20:43:40.809844 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 29 20:43:40 crc kubenswrapper[4741]: I0929 20:43:40.840783 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 20:43:42 crc kubenswrapper[4741]: I0929 20:43:42.187439 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 29 20:43:43 crc kubenswrapper[4741]: I0929 20:43:43.155639 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 20:43:43 crc kubenswrapper[4741]: I0929 20:43:43.155978 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 20:43:43 crc kubenswrapper[4741]: I0929 20:43:43.167484 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 20:43:43 crc kubenswrapper[4741]: I0929 20:43:43.167514 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 20:43:44 crc kubenswrapper[4741]: I0929 20:43:44.283592 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.79:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:43:44 crc kubenswrapper[4741]: I0929 20:43:44.325580 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.78:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:43:44 crc kubenswrapper[4741]: I0929 20:43:44.325705 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.78:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:43:44 crc kubenswrapper[4741]: I0929 20:43:44.325580 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.79:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.294193 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.297120 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.299679 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.317684 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.349227 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.349276 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722344d5-5b89-4178-9187-3d8254c9c11a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.349297 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.349332 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.349353 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2d87\" (UniqueName: \"kubernetes.io/projected/722344d5-5b89-4178-9187-3d8254c9c11a-kube-api-access-k2d87\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.349373 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-scripts\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.450437 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.450706 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722344d5-5b89-4178-9187-3d8254c9c11a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.450804 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.450942 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.451051 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2d87\" (UniqueName: \"kubernetes.io/projected/722344d5-5b89-4178-9187-3d8254c9c11a-kube-api-access-k2d87\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.451153 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-scripts\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.450799 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722344d5-5b89-4178-9187-3d8254c9c11a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.459571 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.465962 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.467307 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-scripts\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.467875 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.474626 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2d87\" (UniqueName: \"kubernetes.io/projected/722344d5-5b89-4178-9187-3d8254c9c11a-kube-api-access-k2d87\") pod \"cinder-scheduler-0\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " pod="openstack/cinder-scheduler-0" Sep 29 20:43:47 crc kubenswrapper[4741]: I0929 20:43:47.623527 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 20:43:48 crc kubenswrapper[4741]: I0929 20:43:48.122111 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:43:48 crc kubenswrapper[4741]: I0929 20:43:48.758250 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:48 crc kubenswrapper[4741]: I0929 20:43:48.758555 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api-log" containerID="cri-o://d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30" gracePeriod=30 Sep 29 20:43:48 crc kubenswrapper[4741]: I0929 20:43:48.758637 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api" containerID="cri-o://ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c" gracePeriod=30 Sep 29 20:43:48 crc kubenswrapper[4741]: I0929 20:43:48.892090 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"722344d5-5b89-4178-9187-3d8254c9c11a","Type":"ContainerStarted","Data":"3a97b7d42e6e4957a2303380143ca96dd446fcccac4b825cf472874fb4561842"} Sep 29 20:43:48 crc kubenswrapper[4741]: I0929 20:43:48.892463 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"722344d5-5b89-4178-9187-3d8254c9c11a","Type":"ContainerStarted","Data":"59f7ba05ca3ffbd36758c7fde30f679d995abb3640701cbc910e18b5a24b636c"} Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.324286 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.326579 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.338664 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.342364 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.488428 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.488492 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.488527 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.488556 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-run\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.488595 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489141 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489264 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4554f451-c9b8-44a6-a5cc-efa6837aa62a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489345 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489407 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489449 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489476 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489658 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489703 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489783 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489817 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtl6n\" (UniqueName: \"kubernetes.io/projected/4554f451-c9b8-44a6-a5cc-efa6837aa62a-kube-api-access-qtl6n\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.489871 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591164 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591212 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591244 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591266 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtl6n\" (UniqueName: \"kubernetes.io/projected/4554f451-c9b8-44a6-a5cc-efa6837aa62a-kube-api-access-qtl6n\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591290 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591327 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591354 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591376 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591413 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-run\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591439 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591455 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591472 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4554f451-c9b8-44a6-a5cc-efa6837aa62a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591494 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591516 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591538 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591557 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.591641 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.592319 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.592347 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.592367 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.592627 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.593449 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-run\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.593535 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.593589 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.593794 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.593942 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4554f451-c9b8-44a6-a5cc-efa6837aa62a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.598032 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.598353 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4554f451-c9b8-44a6-a5cc-efa6837aa62a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.599074 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.599879 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.615138 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4554f451-c9b8-44a6-a5cc-efa6837aa62a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.619040 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtl6n\" (UniqueName: \"kubernetes.io/projected/4554f451-c9b8-44a6-a5cc-efa6837aa62a-kube-api-access-qtl6n\") pod \"cinder-volume-volume1-0\" (UID: \"4554f451-c9b8-44a6-a5cc-efa6837aa62a\") " pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.649973 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.911290 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"722344d5-5b89-4178-9187-3d8254c9c11a","Type":"ContainerStarted","Data":"fd1c2ca8d87cb729195e6ebbe9132f5b9774a6a218b472495babfcd5e1b9b188"} Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.922052 4741 generic.go:334] "Generic (PLEG): container finished" podID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerID="d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30" exitCode=143 Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.922098 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6da521cf-10ef-423e-8649-6452fb96f1d2","Type":"ContainerDied","Data":"d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30"} Sep 29 20:43:49 crc kubenswrapper[4741]: I0929 20:43:49.934265 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.9342368580000002 podStartE2EDuration="2.934236858s" podCreationTimestamp="2025-09-29 20:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:49.929827902 +0000 UTC m=+5671.577617234" watchObservedRunningTime="2025-09-29 20:43:49.934236858 +0000 UTC m=+5671.582026180" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.190086 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Sep 29 20:43:50 crc kubenswrapper[4741]: W0929 20:43:50.200568 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4554f451_c9b8_44a6_a5cc_efa6837aa62a.slice/crio-12eab810496e339edda995e7f7242dbe7a208486099b865cc8d174ff1af75399 WatchSource:0}: Error finding container 12eab810496e339edda995e7f7242dbe7a208486099b865cc8d174ff1af75399: Status 404 returned error can't find the container with id 12eab810496e339edda995e7f7242dbe7a208486099b865cc8d174ff1af75399 Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.202596 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.637485 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.638859 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.646977 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.654809 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.819326 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.819933 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.820098 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-config-data-custom\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.820226 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-nvme\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.820363 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-dev\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.820564 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.820705 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/df7971f7-6713-455c-acad-de1657b4d940-ceph\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.820819 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-sys\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.820928 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-scripts\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.821048 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-lib-modules\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.821168 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.821277 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.821455 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-run\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.821654 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbx58\" (UniqueName: \"kubernetes.io/projected/df7971f7-6713-455c-acad-de1657b4d940-kube-api-access-gbx58\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.821806 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.821936 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-config-data\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.923925 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbx58\" (UniqueName: \"kubernetes.io/projected/df7971f7-6713-455c-acad-de1657b4d940-kube-api-access-gbx58\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.923993 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924030 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-config-data\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924065 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924112 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924132 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-config-data-custom\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924155 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-nvme\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924177 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-dev\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924217 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924238 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-sys\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924257 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/df7971f7-6713-455c-acad-de1657b4d940-ceph\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924283 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-scripts\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924312 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-lib-modules\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924348 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924370 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924445 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-run\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.924548 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-run\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925001 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925801 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-sys\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925882 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-nvme\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925929 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-lib-modules\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925919 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-dev\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925899 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925923 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.925961 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.926006 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/df7971f7-6713-455c-acad-de1657b4d940-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.931614 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.932289 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-scripts\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.935227 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/df7971f7-6713-455c-acad-de1657b4d940-ceph\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.939836 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-config-data\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.942867 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df7971f7-6713-455c-acad-de1657b4d940-config-data-custom\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.943379 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbx58\" (UniqueName: \"kubernetes.io/projected/df7971f7-6713-455c-acad-de1657b4d940-kube-api-access-gbx58\") pod \"cinder-backup-0\" (UID: \"df7971f7-6713-455c-acad-de1657b4d940\") " pod="openstack/cinder-backup-0" Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.943526 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"4554f451-c9b8-44a6-a5cc-efa6837aa62a","Type":"ContainerStarted","Data":"12eab810496e339edda995e7f7242dbe7a208486099b865cc8d174ff1af75399"} Sep 29 20:43:50 crc kubenswrapper[4741]: I0929 20:43:50.960227 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Sep 29 20:43:51 crc kubenswrapper[4741]: W0929 20:43:51.586632 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf7971f7_6713_455c_acad_de1657b4d940.slice/crio-dc0e9016234205f0ee3b18b427dc626e219f26609b5a917b087a49b4fefa2781 WatchSource:0}: Error finding container dc0e9016234205f0ee3b18b427dc626e219f26609b5a917b087a49b4fefa2781: Status 404 returned error can't find the container with id dc0e9016234205f0ee3b18b427dc626e219f26609b5a917b087a49b4fefa2781 Sep 29 20:43:51 crc kubenswrapper[4741]: I0929 20:43:51.587420 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Sep 29 20:43:51 crc kubenswrapper[4741]: I0929 20:43:51.925500 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.74:8776/healthcheck\": read tcp 10.217.0.2:44700->10.217.1.74:8776: read: connection reset by peer" Sep 29 20:43:51 crc kubenswrapper[4741]: I0929 20:43:51.952437 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"df7971f7-6713-455c-acad-de1657b4d940","Type":"ContainerStarted","Data":"dc0e9016234205f0ee3b18b427dc626e219f26609b5a917b087a49b4fefa2781"} Sep 29 20:43:51 crc kubenswrapper[4741]: I0929 20:43:51.954556 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"4554f451-c9b8-44a6-a5cc-efa6837aa62a","Type":"ContainerStarted","Data":"86809f86e07d025f608a9e782e688c22ea593d5d8d9830878934cdc08be9525b"} Sep 29 20:43:51 crc kubenswrapper[4741]: I0929 20:43:51.954601 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"4554f451-c9b8-44a6-a5cc-efa6837aa62a","Type":"ContainerStarted","Data":"e6795d8729bf3dfb434b5bb3e8976ab5229c348a752106b19ef56fbc6748f732"} Sep 29 20:43:51 crc kubenswrapper[4741]: I0929 20:43:51.975606 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.046861578 podStartE2EDuration="2.975589897s" podCreationTimestamp="2025-09-29 20:43:49 +0000 UTC" firstStartedPulling="2025-09-29 20:43:50.202195632 +0000 UTC m=+5671.849984994" lastFinishedPulling="2025-09-29 20:43:51.130923981 +0000 UTC m=+5672.778713313" observedRunningTime="2025-09-29 20:43:51.972242083 +0000 UTC m=+5673.620031415" watchObservedRunningTime="2025-09-29 20:43:51.975589897 +0000 UTC m=+5673.623379229" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.324159 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.454887 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-combined-ca-bundle\") pod \"6da521cf-10ef-423e-8649-6452fb96f1d2\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.454994 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6da521cf-10ef-423e-8649-6452fb96f1d2-logs\") pod \"6da521cf-10ef-423e-8649-6452fb96f1d2\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.455017 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data-custom\") pod \"6da521cf-10ef-423e-8649-6452fb96f1d2\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.455068 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data\") pod \"6da521cf-10ef-423e-8649-6452fb96f1d2\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.455119 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-scripts\") pod \"6da521cf-10ef-423e-8649-6452fb96f1d2\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.455166 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9mlh\" (UniqueName: \"kubernetes.io/projected/6da521cf-10ef-423e-8649-6452fb96f1d2-kube-api-access-c9mlh\") pod \"6da521cf-10ef-423e-8649-6452fb96f1d2\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.455271 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da521cf-10ef-423e-8649-6452fb96f1d2-etc-machine-id\") pod \"6da521cf-10ef-423e-8649-6452fb96f1d2\" (UID: \"6da521cf-10ef-423e-8649-6452fb96f1d2\") " Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.456051 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6da521cf-10ef-423e-8649-6452fb96f1d2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6da521cf-10ef-423e-8649-6452fb96f1d2" (UID: "6da521cf-10ef-423e-8649-6452fb96f1d2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.456719 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6da521cf-10ef-423e-8649-6452fb96f1d2-logs" (OuterVolumeSpecName: "logs") pod "6da521cf-10ef-423e-8649-6452fb96f1d2" (UID: "6da521cf-10ef-423e-8649-6452fb96f1d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.462485 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6da521cf-10ef-423e-8649-6452fb96f1d2" (UID: "6da521cf-10ef-423e-8649-6452fb96f1d2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.463072 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-scripts" (OuterVolumeSpecName: "scripts") pod "6da521cf-10ef-423e-8649-6452fb96f1d2" (UID: "6da521cf-10ef-423e-8649-6452fb96f1d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.484790 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da521cf-10ef-423e-8649-6452fb96f1d2-kube-api-access-c9mlh" (OuterVolumeSpecName: "kube-api-access-c9mlh") pod "6da521cf-10ef-423e-8649-6452fb96f1d2" (UID: "6da521cf-10ef-423e-8649-6452fb96f1d2"). InnerVolumeSpecName "kube-api-access-c9mlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.517413 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6da521cf-10ef-423e-8649-6452fb96f1d2" (UID: "6da521cf-10ef-423e-8649-6452fb96f1d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.557123 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.557163 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6da521cf-10ef-423e-8649-6452fb96f1d2-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.557179 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.557191 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9mlh\" (UniqueName: \"kubernetes.io/projected/6da521cf-10ef-423e-8649-6452fb96f1d2-kube-api-access-c9mlh\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.557204 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da521cf-10ef-423e-8649-6452fb96f1d2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.557214 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.569476 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data" (OuterVolumeSpecName: "config-data") pod "6da521cf-10ef-423e-8649-6452fb96f1d2" (UID: "6da521cf-10ef-423e-8649-6452fb96f1d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.624023 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.658984 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da521cf-10ef-423e-8649-6452fb96f1d2-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.963680 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"df7971f7-6713-455c-acad-de1657b4d940","Type":"ContainerStarted","Data":"e74594062c3e1cc01cdc7868413c65e0b22532bfef40b6dfe1aae4b1b513bc44"} Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.963996 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"df7971f7-6713-455c-acad-de1657b4d940","Type":"ContainerStarted","Data":"57270d788f3c79e704240841645750674de431206348f7af6e7e305889d65878"} Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.968304 4741 generic.go:334] "Generic (PLEG): container finished" podID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerID="ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c" exitCode=0 Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.968484 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6da521cf-10ef-423e-8649-6452fb96f1d2","Type":"ContainerDied","Data":"ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c"} Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.968511 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6da521cf-10ef-423e-8649-6452fb96f1d2","Type":"ContainerDied","Data":"f0f9d90d447776c3f298c647a65bebc73a432956b5e66f33bb47c2094794135b"} Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.968529 4741 scope.go:117] "RemoveContainer" containerID="ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.968647 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 20:43:52 crc kubenswrapper[4741]: I0929 20:43:52.996538 4741 scope.go:117] "RemoveContainer" containerID="d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.003197 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.042050954 podStartE2EDuration="3.003186576s" podCreationTimestamp="2025-09-29 20:43:50 +0000 UTC" firstStartedPulling="2025-09-29 20:43:51.588649819 +0000 UTC m=+5673.236439141" lastFinishedPulling="2025-09-29 20:43:52.549785431 +0000 UTC m=+5674.197574763" observedRunningTime="2025-09-29 20:43:52.997616883 +0000 UTC m=+5674.645406215" watchObservedRunningTime="2025-09-29 20:43:53.003186576 +0000 UTC m=+5674.650975908" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.049136 4741 scope.go:117] "RemoveContainer" containerID="ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c" Sep 29 20:43:53 crc kubenswrapper[4741]: E0929 20:43:53.049602 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c\": container with ID starting with ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c not found: ID does not exist" containerID="ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.049647 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c"} err="failed to get container status \"ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c\": rpc error: code = NotFound desc = could not find container \"ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c\": container with ID starting with ca761c93569cb625f309245168545201682d8bcaf4168282cc1a7f2b588bc89c not found: ID does not exist" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.049679 4741 scope.go:117] "RemoveContainer" containerID="d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30" Sep 29 20:43:53 crc kubenswrapper[4741]: E0929 20:43:53.049986 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30\": container with ID starting with d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30 not found: ID does not exist" containerID="d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.050018 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30"} err="failed to get container status \"d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30\": rpc error: code = NotFound desc = could not find container \"d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30\": container with ID starting with d882d9cfbab125ba3647eec76b8453c3b9ed06dca6012f58d56fe586056c3e30 not found: ID does not exist" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.083692 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.098799 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.106282 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:53 crc kubenswrapper[4741]: E0929 20:43:53.106796 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api-log" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.106814 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api-log" Sep 29 20:43:53 crc kubenswrapper[4741]: E0929 20:43:53.106829 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.106838 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.107105 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.107128 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" containerName="cinder-api-log" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.108488 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.110491 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.119015 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.160012 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.160526 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.162129 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.162279 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.169765 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.169860 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.174269 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.176597 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.180733 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.270312 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbnvw\" (UniqueName: \"kubernetes.io/projected/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-kube-api-access-cbnvw\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.270423 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.270450 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-config-data\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.270580 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-config-data-custom\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.270693 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.270812 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-scripts\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.270864 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-logs\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.372464 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.372541 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-scripts\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.372568 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-logs\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.372640 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbnvw\" (UniqueName: \"kubernetes.io/projected/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-kube-api-access-cbnvw\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.372703 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.372726 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-config-data\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.372757 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-config-data-custom\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.373048 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-logs\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.373965 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.379437 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-scripts\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.379596 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.379927 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-config-data-custom\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.384286 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-config-data\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.392293 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbnvw\" (UniqueName: \"kubernetes.io/projected/2bbfd908-1ced-4cae-8f3d-d7c89ec47c15-kube-api-access-cbnvw\") pod \"cinder-api-0\" (UID: \"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15\") " pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.436383 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.920195 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 20:43:53 crc kubenswrapper[4741]: W0929 20:43:53.923003 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bbfd908_1ced_4cae_8f3d_d7c89ec47c15.slice/crio-e470ff82a97f2d404ff6bb78aa2544ab3e47940e21994e657ce743783b1dcfb5 WatchSource:0}: Error finding container e470ff82a97f2d404ff6bb78aa2544ab3e47940e21994e657ce743783b1dcfb5: Status 404 returned error can't find the container with id e470ff82a97f2d404ff6bb78aa2544ab3e47940e21994e657ce743783b1dcfb5 Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.984381 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15","Type":"ContainerStarted","Data":"e470ff82a97f2d404ff6bb78aa2544ab3e47940e21994e657ce743783b1dcfb5"} Sep 29 20:43:53 crc kubenswrapper[4741]: I0929 20:43:53.987260 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 20:43:54 crc kubenswrapper[4741]: I0929 20:43:54.660008 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Sep 29 20:43:54 crc kubenswrapper[4741]: I0929 20:43:54.996479 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15","Type":"ContainerStarted","Data":"63429e6aa659a4f8631f08e9e4acfab519f156c0364b37a975e05d306f715395"} Sep 29 20:43:55 crc kubenswrapper[4741]: I0929 20:43:55.100364 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da521cf-10ef-423e-8649-6452fb96f1d2" path="/var/lib/kubelet/pods/6da521cf-10ef-423e-8649-6452fb96f1d2/volumes" Sep 29 20:43:55 crc kubenswrapper[4741]: I0929 20:43:55.961080 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Sep 29 20:43:56 crc kubenswrapper[4741]: I0929 20:43:56.007688 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2bbfd908-1ced-4cae-8f3d-d7c89ec47c15","Type":"ContainerStarted","Data":"a40d61b34aba9364cfa82103d724ea089681429ef2d2df5e518f9453ba438a35"} Sep 29 20:43:56 crc kubenswrapper[4741]: I0929 20:43:56.007895 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 29 20:43:56 crc kubenswrapper[4741]: I0929 20:43:56.050905 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.050886145 podStartE2EDuration="3.050886145s" podCreationTimestamp="2025-09-29 20:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:43:56.039090921 +0000 UTC m=+5677.686880263" watchObservedRunningTime="2025-09-29 20:43:56.050886145 +0000 UTC m=+5677.698675477" Sep 29 20:43:57 crc kubenswrapper[4741]: I0929 20:43:57.975214 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 29 20:43:58 crc kubenswrapper[4741]: I0929 20:43:58.087219 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:43:58 crc kubenswrapper[4741]: I0929 20:43:58.087725 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="cinder-scheduler" containerID="cri-o://3a97b7d42e6e4957a2303380143ca96dd446fcccac4b825cf472874fb4561842" gracePeriod=30 Sep 29 20:43:58 crc kubenswrapper[4741]: I0929 20:43:58.087991 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="probe" containerID="cri-o://fd1c2ca8d87cb729195e6ebbe9132f5b9774a6a218b472495babfcd5e1b9b188" gracePeriod=30 Sep 29 20:43:59 crc kubenswrapper[4741]: I0929 20:43:59.047194 4741 generic.go:334] "Generic (PLEG): container finished" podID="722344d5-5b89-4178-9187-3d8254c9c11a" containerID="fd1c2ca8d87cb729195e6ebbe9132f5b9774a6a218b472495babfcd5e1b9b188" exitCode=0 Sep 29 20:43:59 crc kubenswrapper[4741]: I0929 20:43:59.047252 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"722344d5-5b89-4178-9187-3d8254c9c11a","Type":"ContainerDied","Data":"fd1c2ca8d87cb729195e6ebbe9132f5b9774a6a218b472495babfcd5e1b9b188"} Sep 29 20:43:59 crc kubenswrapper[4741]: I0929 20:43:59.873901 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.059284 4741 generic.go:334] "Generic (PLEG): container finished" podID="722344d5-5b89-4178-9187-3d8254c9c11a" containerID="3a97b7d42e6e4957a2303380143ca96dd446fcccac4b825cf472874fb4561842" exitCode=0 Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.059325 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"722344d5-5b89-4178-9187-3d8254c9c11a","Type":"ContainerDied","Data":"3a97b7d42e6e4957a2303380143ca96dd446fcccac4b825cf472874fb4561842"} Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.198948 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298282 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2d87\" (UniqueName: \"kubernetes.io/projected/722344d5-5b89-4178-9187-3d8254c9c11a-kube-api-access-k2d87\") pod \"722344d5-5b89-4178-9187-3d8254c9c11a\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298350 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-scripts\") pod \"722344d5-5b89-4178-9187-3d8254c9c11a\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298379 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data-custom\") pod \"722344d5-5b89-4178-9187-3d8254c9c11a\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298421 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-combined-ca-bundle\") pod \"722344d5-5b89-4178-9187-3d8254c9c11a\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298478 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722344d5-5b89-4178-9187-3d8254c9c11a-etc-machine-id\") pod \"722344d5-5b89-4178-9187-3d8254c9c11a\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298601 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data\") pod \"722344d5-5b89-4178-9187-3d8254c9c11a\" (UID: \"722344d5-5b89-4178-9187-3d8254c9c11a\") " Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298673 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722344d5-5b89-4178-9187-3d8254c9c11a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "722344d5-5b89-4178-9187-3d8254c9c11a" (UID: "722344d5-5b89-4178-9187-3d8254c9c11a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.298945 4741 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722344d5-5b89-4178-9187-3d8254c9c11a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.303339 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "722344d5-5b89-4178-9187-3d8254c9c11a" (UID: "722344d5-5b89-4178-9187-3d8254c9c11a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.303548 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/722344d5-5b89-4178-9187-3d8254c9c11a-kube-api-access-k2d87" (OuterVolumeSpecName: "kube-api-access-k2d87") pod "722344d5-5b89-4178-9187-3d8254c9c11a" (UID: "722344d5-5b89-4178-9187-3d8254c9c11a"). InnerVolumeSpecName "kube-api-access-k2d87". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.311560 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-scripts" (OuterVolumeSpecName: "scripts") pod "722344d5-5b89-4178-9187-3d8254c9c11a" (UID: "722344d5-5b89-4178-9187-3d8254c9c11a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.348245 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "722344d5-5b89-4178-9187-3d8254c9c11a" (UID: "722344d5-5b89-4178-9187-3d8254c9c11a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.394481 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data" (OuterVolumeSpecName: "config-data") pod "722344d5-5b89-4178-9187-3d8254c9c11a" (UID: "722344d5-5b89-4178-9187-3d8254c9c11a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.401078 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.401109 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2d87\" (UniqueName: \"kubernetes.io/projected/722344d5-5b89-4178-9187-3d8254c9c11a-kube-api-access-k2d87\") on node \"crc\" DevicePath \"\"" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.401135 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.401147 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 20:44:00 crc kubenswrapper[4741]: I0929 20:44:00.401156 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722344d5-5b89-4178-9187-3d8254c9c11a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.070865 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"722344d5-5b89-4178-9187-3d8254c9c11a","Type":"ContainerDied","Data":"59f7ba05ca3ffbd36758c7fde30f679d995abb3640701cbc910e18b5a24b636c"} Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.070929 4741 scope.go:117] "RemoveContainer" containerID="fd1c2ca8d87cb729195e6ebbe9132f5b9774a6a218b472495babfcd5e1b9b188" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.070987 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.107628 4741 scope.go:117] "RemoveContainer" containerID="3a97b7d42e6e4957a2303380143ca96dd446fcccac4b825cf472874fb4561842" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.115203 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.137465 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.164472 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:44:01 crc kubenswrapper[4741]: E0929 20:44:01.164928 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="probe" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.164947 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="probe" Sep 29 20:44:01 crc kubenswrapper[4741]: E0929 20:44:01.164977 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="cinder-scheduler" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.164983 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="cinder-scheduler" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.165155 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="cinder-scheduler" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.165186 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" containerName="probe" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.166251 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.168024 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.177309 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.232054 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.317059 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nvmv\" (UniqueName: \"kubernetes.io/projected/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-kube-api-access-8nvmv\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.317108 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.317176 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.317219 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.317235 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-config-data\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.317299 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-scripts\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.419443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nvmv\" (UniqueName: \"kubernetes.io/projected/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-kube-api-access-8nvmv\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.419772 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.419919 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.420056 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.420128 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.420143 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-config-data\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.420334 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-scripts\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.424687 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.426467 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.427106 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-scripts\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.427295 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-config-data\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.439617 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nvmv\" (UniqueName: \"kubernetes.io/projected/ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04-kube-api-access-8nvmv\") pod \"cinder-scheduler-0\" (UID: \"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04\") " pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.489156 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.738787 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.738834 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:44:01 crc kubenswrapper[4741]: I0929 20:44:01.911409 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 20:44:01 crc kubenswrapper[4741]: W0929 20:44:01.918434 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba97b0c5_5fa9_42b4_bb96_a02a5c8f8b04.slice/crio-01bc1f81c0c0f9512684383c07c094318aaefc9f2b17f61fb27960ade51c74f5 WatchSource:0}: Error finding container 01bc1f81c0c0f9512684383c07c094318aaefc9f2b17f61fb27960ade51c74f5: Status 404 returned error can't find the container with id 01bc1f81c0c0f9512684383c07c094318aaefc9f2b17f61fb27960ade51c74f5 Sep 29 20:44:02 crc kubenswrapper[4741]: I0929 20:44:02.082237 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04","Type":"ContainerStarted","Data":"01bc1f81c0c0f9512684383c07c094318aaefc9f2b17f61fb27960ade51c74f5"} Sep 29 20:44:03 crc kubenswrapper[4741]: I0929 20:44:03.134265 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="722344d5-5b89-4178-9187-3d8254c9c11a" path="/var/lib/kubelet/pods/722344d5-5b89-4178-9187-3d8254c9c11a/volumes" Sep 29 20:44:03 crc kubenswrapper[4741]: I0929 20:44:03.140868 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04","Type":"ContainerStarted","Data":"f0419613a11da21697e2c04921d5c064900019664b8e58f349589d0e4bf49431"} Sep 29 20:44:03 crc kubenswrapper[4741]: I0929 20:44:03.141097 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04","Type":"ContainerStarted","Data":"14b96383f20f470d2d70c1da5ab9c7af9eb56ff35d587401354763ced0c07699"} Sep 29 20:44:05 crc kubenswrapper[4741]: I0929 20:44:05.197933 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 29 20:44:05 crc kubenswrapper[4741]: I0929 20:44:05.231522 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.231497335 podStartE2EDuration="4.231497335s" podCreationTimestamp="2025-09-29 20:44:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:44:03.149915441 +0000 UTC m=+5684.797704773" watchObservedRunningTime="2025-09-29 20:44:05.231497335 +0000 UTC m=+5686.879286667" Sep 29 20:44:06 crc kubenswrapper[4741]: I0929 20:44:06.489821 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 29 20:44:11 crc kubenswrapper[4741]: I0929 20:44:11.673714 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 29 20:44:29 crc kubenswrapper[4741]: I0929 20:44:29.049625 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-pgfbv"] Sep 29 20:44:29 crc kubenswrapper[4741]: I0929 20:44:29.056635 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-pgfbv"] Sep 29 20:44:29 crc kubenswrapper[4741]: I0929 20:44:29.096213 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08f96dea-793e-46e3-9249-033ba38c6607" path="/var/lib/kubelet/pods/08f96dea-793e-46e3-9249-033ba38c6607/volumes" Sep 29 20:44:31 crc kubenswrapper[4741]: I0929 20:44:31.739584 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:44:31 crc kubenswrapper[4741]: I0929 20:44:31.740022 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:44:39 crc kubenswrapper[4741]: I0929 20:44:39.041121 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9b22-account-create-55fn4"] Sep 29 20:44:39 crc kubenswrapper[4741]: I0929 20:44:39.059018 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9b22-account-create-55fn4"] Sep 29 20:44:39 crc kubenswrapper[4741]: I0929 20:44:39.096511 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afefc204-e529-41d6-9592-22c926ccbf25" path="/var/lib/kubelet/pods/afefc204-e529-41d6-9592-22c926ccbf25/volumes" Sep 29 20:44:45 crc kubenswrapper[4741]: I0929 20:44:45.037242 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-qbkn6"] Sep 29 20:44:45 crc kubenswrapper[4741]: I0929 20:44:45.047745 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-qbkn6"] Sep 29 20:44:45 crc kubenswrapper[4741]: I0929 20:44:45.101674 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b16a61e9-d870-4046-8cda-25d5980fb655" path="/var/lib/kubelet/pods/b16a61e9-d870-4046-8cda-25d5980fb655/volumes" Sep 29 20:44:59 crc kubenswrapper[4741]: I0929 20:44:59.026269 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nkcj5"] Sep 29 20:44:59 crc kubenswrapper[4741]: I0929 20:44:59.035218 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nkcj5"] Sep 29 20:44:59 crc kubenswrapper[4741]: I0929 20:44:59.099633 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8650376-c59a-47cd-9ff6-a688a16516ed" path="/var/lib/kubelet/pods/a8650376-c59a-47cd-9ff6-a688a16516ed/volumes" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.156010 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj"] Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.158327 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.161626 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.162627 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.185003 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj"] Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.279132 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b60a4868-d1e2-495a-9072-804ba359f428-config-volume\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.279211 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b60a4868-d1e2-495a-9072-804ba359f428-secret-volume\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.279242 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt452\" (UniqueName: \"kubernetes.io/projected/b60a4868-d1e2-495a-9072-804ba359f428-kube-api-access-dt452\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.380697 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b60a4868-d1e2-495a-9072-804ba359f428-config-volume\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.380772 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b60a4868-d1e2-495a-9072-804ba359f428-secret-volume\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.380794 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt452\" (UniqueName: \"kubernetes.io/projected/b60a4868-d1e2-495a-9072-804ba359f428-kube-api-access-dt452\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.381760 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b60a4868-d1e2-495a-9072-804ba359f428-config-volume\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.394503 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b60a4868-d1e2-495a-9072-804ba359f428-secret-volume\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.395865 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt452\" (UniqueName: \"kubernetes.io/projected/b60a4868-d1e2-495a-9072-804ba359f428-kube-api-access-dt452\") pod \"collect-profiles-29319645-2rvlj\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.494231 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:00 crc kubenswrapper[4741]: I0929 20:45:00.937004 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj"] Sep 29 20:45:00 crc kubenswrapper[4741]: W0929 20:45:00.941445 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb60a4868_d1e2_495a_9072_804ba359f428.slice/crio-520cd46a535a6093304b8ecfa4dafb8c6763eebd0dd87c37c2612a10832686a1 WatchSource:0}: Error finding container 520cd46a535a6093304b8ecfa4dafb8c6763eebd0dd87c37c2612a10832686a1: Status 404 returned error can't find the container with id 520cd46a535a6093304b8ecfa4dafb8c6763eebd0dd87c37c2612a10832686a1 Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.717658 4741 generic.go:334] "Generic (PLEG): container finished" podID="b60a4868-d1e2-495a-9072-804ba359f428" containerID="3ac79bbaa6afbb06d0adb6779f2fcaed485593c93f1cc1100dafa2950133b104" exitCode=0 Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.717803 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" event={"ID":"b60a4868-d1e2-495a-9072-804ba359f428","Type":"ContainerDied","Data":"3ac79bbaa6afbb06d0adb6779f2fcaed485593c93f1cc1100dafa2950133b104"} Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.718015 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" event={"ID":"b60a4868-d1e2-495a-9072-804ba359f428","Type":"ContainerStarted","Data":"520cd46a535a6093304b8ecfa4dafb8c6763eebd0dd87c37c2612a10832686a1"} Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.739886 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.739978 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.740047 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.741382 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"564907e3fd83739b2ba2e94f48864f64d35270919df4a735eaebeff724bee01a"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:45:01 crc kubenswrapper[4741]: I0929 20:45:01.741496 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://564907e3fd83739b2ba2e94f48864f64d35270919df4a735eaebeff724bee01a" gracePeriod=600 Sep 29 20:45:02 crc kubenswrapper[4741]: I0929 20:45:02.732526 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="564907e3fd83739b2ba2e94f48864f64d35270919df4a735eaebeff724bee01a" exitCode=0 Sep 29 20:45:02 crc kubenswrapper[4741]: I0929 20:45:02.732641 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"564907e3fd83739b2ba2e94f48864f64d35270919df4a735eaebeff724bee01a"} Sep 29 20:45:02 crc kubenswrapper[4741]: I0929 20:45:02.733969 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7"} Sep 29 20:45:02 crc kubenswrapper[4741]: I0929 20:45:02.734016 4741 scope.go:117] "RemoveContainer" containerID="0e054d87c4b9a9dd9d7691d35999ce9b8331e73c7d4d434e8c89ce160fde5ec4" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.066419 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.136222 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b60a4868-d1e2-495a-9072-804ba359f428-secret-volume\") pod \"b60a4868-d1e2-495a-9072-804ba359f428\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.136325 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt452\" (UniqueName: \"kubernetes.io/projected/b60a4868-d1e2-495a-9072-804ba359f428-kube-api-access-dt452\") pod \"b60a4868-d1e2-495a-9072-804ba359f428\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.136380 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b60a4868-d1e2-495a-9072-804ba359f428-config-volume\") pod \"b60a4868-d1e2-495a-9072-804ba359f428\" (UID: \"b60a4868-d1e2-495a-9072-804ba359f428\") " Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.137160 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b60a4868-d1e2-495a-9072-804ba359f428-config-volume" (OuterVolumeSpecName: "config-volume") pod "b60a4868-d1e2-495a-9072-804ba359f428" (UID: "b60a4868-d1e2-495a-9072-804ba359f428"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.137754 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b60a4868-d1e2-495a-9072-804ba359f428-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.141882 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b60a4868-d1e2-495a-9072-804ba359f428-kube-api-access-dt452" (OuterVolumeSpecName: "kube-api-access-dt452") pod "b60a4868-d1e2-495a-9072-804ba359f428" (UID: "b60a4868-d1e2-495a-9072-804ba359f428"). InnerVolumeSpecName "kube-api-access-dt452". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.143095 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b60a4868-d1e2-495a-9072-804ba359f428-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b60a4868-d1e2-495a-9072-804ba359f428" (UID: "b60a4868-d1e2-495a-9072-804ba359f428"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.239692 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b60a4868-d1e2-495a-9072-804ba359f428-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.239736 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt452\" (UniqueName: \"kubernetes.io/projected/b60a4868-d1e2-495a-9072-804ba359f428-kube-api-access-dt452\") on node \"crc\" DevicePath \"\"" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.749637 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" event={"ID":"b60a4868-d1e2-495a-9072-804ba359f428","Type":"ContainerDied","Data":"520cd46a535a6093304b8ecfa4dafb8c6763eebd0dd87c37c2612a10832686a1"} Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.750080 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="520cd46a535a6093304b8ecfa4dafb8c6763eebd0dd87c37c2612a10832686a1" Sep 29 20:45:03 crc kubenswrapper[4741]: I0929 20:45:03.749668 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj" Sep 29 20:45:04 crc kubenswrapper[4741]: I0929 20:45:04.140347 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg"] Sep 29 20:45:04 crc kubenswrapper[4741]: I0929 20:45:04.149632 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319600-8fhqg"] Sep 29 20:45:05 crc kubenswrapper[4741]: I0929 20:45:05.100612 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9a64fc-57a2-49a6-bccc-25a23ae2835a" path="/var/lib/kubelet/pods/1c9a64fc-57a2-49a6-bccc-25a23ae2835a/volumes" Sep 29 20:45:27 crc kubenswrapper[4741]: I0929 20:45:27.582977 4741 scope.go:117] "RemoveContainer" containerID="64c212e1a8175794b79a0d1c3960b1990d47cdf75ba74c0bdffd833d8fdac26c" Sep 29 20:45:27 crc kubenswrapper[4741]: I0929 20:45:27.634489 4741 scope.go:117] "RemoveContainer" containerID="3430ec746ab2a3c23749a224cdb681191d6b55f8b86d5c8b765209f08709bde6" Sep 29 20:45:27 crc kubenswrapper[4741]: I0929 20:45:27.724004 4741 scope.go:117] "RemoveContainer" containerID="0dc0f530e90466049373310a4126962ad5086cc3a72062fe81b0e165d9ca10a1" Sep 29 20:45:27 crc kubenswrapper[4741]: I0929 20:45:27.748289 4741 scope.go:117] "RemoveContainer" containerID="a32763cb55fdc7bb481856264c284ca1ec29d9aa5fc4b5d51a06e1b21a9454d3" Sep 29 20:45:27 crc kubenswrapper[4741]: I0929 20:45:27.794100 4741 scope.go:117] "RemoveContainer" containerID="4d82aae8ccf2e5120dc9475a95d58491a15b05a856d163470bcbe7c53bb56db6" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.873674 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xbqvm"] Sep 29 20:45:54 crc kubenswrapper[4741]: E0929 20:45:54.874560 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b60a4868-d1e2-495a-9072-804ba359f428" containerName="collect-profiles" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.874575 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b60a4868-d1e2-495a-9072-804ba359f428" containerName="collect-profiles" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.874777 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b60a4868-d1e2-495a-9072-804ba359f428" containerName="collect-profiles" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.875450 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.878070 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.884062 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-9hxqh" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.898878 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-mlwzz"] Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.901367 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.911174 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xbqvm"] Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.922198 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mlwzz"] Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998380 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8755cb82-1296-4625-8767-4033013502ca-scripts\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998449 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-run\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998479 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhzrh\" (UniqueName: \"kubernetes.io/projected/32d7ebdb-b4ae-4d70-ace7-458c7334b352-kube-api-access-fhzrh\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998631 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntqx9\" (UniqueName: \"kubernetes.io/projected/8755cb82-1296-4625-8767-4033013502ca-kube-api-access-ntqx9\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998716 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-lib\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998766 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32d7ebdb-b4ae-4d70-ace7-458c7334b352-scripts\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998818 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-log\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998840 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-etc-ovs\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998866 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-run-ovn\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998918 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-log-ovn\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:54 crc kubenswrapper[4741]: I0929 20:45:54.998972 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-run\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100673 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntqx9\" (UniqueName: \"kubernetes.io/projected/8755cb82-1296-4625-8767-4033013502ca-kube-api-access-ntqx9\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100714 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-lib\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100738 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32d7ebdb-b4ae-4d70-ace7-458c7334b352-scripts\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100775 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-log\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100794 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-etc-ovs\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100818 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-run-ovn\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100843 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-log-ovn\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100860 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-run\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100897 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8755cb82-1296-4625-8767-4033013502ca-scripts\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100920 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-run\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.100944 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhzrh\" (UniqueName: \"kubernetes.io/projected/32d7ebdb-b4ae-4d70-ace7-458c7334b352-kube-api-access-fhzrh\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.101411 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-run-ovn\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.101464 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-log-ovn\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.101487 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-lib\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.101530 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8755cb82-1296-4625-8767-4033013502ca-var-run\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.101794 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-run\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.101880 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-var-log\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.101862 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/32d7ebdb-b4ae-4d70-ace7-458c7334b352-etc-ovs\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.103461 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32d7ebdb-b4ae-4d70-ace7-458c7334b352-scripts\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.103483 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8755cb82-1296-4625-8767-4033013502ca-scripts\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.123098 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhzrh\" (UniqueName: \"kubernetes.io/projected/32d7ebdb-b4ae-4d70-ace7-458c7334b352-kube-api-access-fhzrh\") pod \"ovn-controller-ovs-mlwzz\" (UID: \"32d7ebdb-b4ae-4d70-ace7-458c7334b352\") " pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.123357 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntqx9\" (UniqueName: \"kubernetes.io/projected/8755cb82-1296-4625-8767-4033013502ca-kube-api-access-ntqx9\") pod \"ovn-controller-xbqvm\" (UID: \"8755cb82-1296-4625-8767-4033013502ca\") " pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.192926 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.227317 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:55 crc kubenswrapper[4741]: I0929 20:45:55.718979 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xbqvm"] Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.048377 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mlwzz"] Sep 29 20:45:56 crc kubenswrapper[4741]: W0929 20:45:56.055076 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32d7ebdb_b4ae_4d70_ace7_458c7334b352.slice/crio-f4c8953687e9017987b94e7363d80b80893fdc6313ca7b6d3a3dff8fdecf4131 WatchSource:0}: Error finding container f4c8953687e9017987b94e7363d80b80893fdc6313ca7b6d3a3dff8fdecf4131: Status 404 returned error can't find the container with id f4c8953687e9017987b94e7363d80b80893fdc6313ca7b6d3a3dff8fdecf4131 Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.272176 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm" event={"ID":"8755cb82-1296-4625-8767-4033013502ca","Type":"ContainerStarted","Data":"ff9a4151c905d13179675e4ffcf43d6b161f7d5d7400496a8c253f05bd642328"} Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.272479 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm" event={"ID":"8755cb82-1296-4625-8767-4033013502ca","Type":"ContainerStarted","Data":"eaa6e55a3bb19cbe435337aab65a9fc63d99edff2aaf9d05030e961bfd6a23ea"} Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.274022 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-xbqvm" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.275459 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mlwzz" event={"ID":"32d7ebdb-b4ae-4d70-ace7-458c7334b352","Type":"ContainerStarted","Data":"f4c8953687e9017987b94e7363d80b80893fdc6313ca7b6d3a3dff8fdecf4131"} Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.298509 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xbqvm" podStartSLOduration=2.298490619 podStartE2EDuration="2.298490619s" podCreationTimestamp="2025-09-29 20:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:45:56.298448488 +0000 UTC m=+5797.946237820" watchObservedRunningTime="2025-09-29 20:45:56.298490619 +0000 UTC m=+5797.946279941" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.381575 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-gl88h"] Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.383435 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.385380 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.388931 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gl88h"] Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.532001 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-ovn-rundir\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.532069 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-config\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.532274 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zxfc\" (UniqueName: \"kubernetes.io/projected/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-kube-api-access-8zxfc\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.532479 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-ovs-rundir\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.633858 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-config\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.634033 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zxfc\" (UniqueName: \"kubernetes.io/projected/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-kube-api-access-8zxfc\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.634088 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-ovs-rundir\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.634172 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-ovn-rundir\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.634491 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-ovn-rundir\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.634548 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-ovs-rundir\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.634592 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-config\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.659594 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zxfc\" (UniqueName: \"kubernetes.io/projected/9cc73ea6-7c30-4a29-be40-7b85ceaeff71-kube-api-access-8zxfc\") pod \"ovn-controller-metrics-gl88h\" (UID: \"9cc73ea6-7c30-4a29-be40-7b85ceaeff71\") " pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:56 crc kubenswrapper[4741]: I0929 20:45:56.740214 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gl88h" Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.164714 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gl88h"] Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.285403 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gl88h" event={"ID":"9cc73ea6-7c30-4a29-be40-7b85ceaeff71","Type":"ContainerStarted","Data":"8552ae3c398e844163f62dc8c8ea16b258d777564f424785da64f749172ef9ec"} Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.286818 4741 generic.go:334] "Generic (PLEG): container finished" podID="32d7ebdb-b4ae-4d70-ace7-458c7334b352" containerID="5a96704029afb506f0e87f18f74d30e1b346889d035df1ce07eeea6235e008c4" exitCode=0 Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.286878 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mlwzz" event={"ID":"32d7ebdb-b4ae-4d70-ace7-458c7334b352","Type":"ContainerDied","Data":"5a96704029afb506f0e87f18f74d30e1b346889d035df1ce07eeea6235e008c4"} Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.657844 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-fsrvf"] Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.662985 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fsrvf" Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.674145 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-fsrvf"] Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.756545 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glc8z\" (UniqueName: \"kubernetes.io/projected/deda0fe8-0c19-4799-a2de-45c6a10014cb-kube-api-access-glc8z\") pod \"octavia-db-create-fsrvf\" (UID: \"deda0fe8-0c19-4799-a2de-45c6a10014cb\") " pod="openstack/octavia-db-create-fsrvf" Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.858341 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glc8z\" (UniqueName: \"kubernetes.io/projected/deda0fe8-0c19-4799-a2de-45c6a10014cb-kube-api-access-glc8z\") pod \"octavia-db-create-fsrvf\" (UID: \"deda0fe8-0c19-4799-a2de-45c6a10014cb\") " pod="openstack/octavia-db-create-fsrvf" Sep 29 20:45:57 crc kubenswrapper[4741]: I0929 20:45:57.881249 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glc8z\" (UniqueName: \"kubernetes.io/projected/deda0fe8-0c19-4799-a2de-45c6a10014cb-kube-api-access-glc8z\") pod \"octavia-db-create-fsrvf\" (UID: \"deda0fe8-0c19-4799-a2de-45c6a10014cb\") " pod="openstack/octavia-db-create-fsrvf" Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.002849 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fsrvf" Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.299171 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mlwzz" event={"ID":"32d7ebdb-b4ae-4d70-ace7-458c7334b352","Type":"ContainerStarted","Data":"754a765715fb867e91c7e6430659316f5cf94dc24ca19c0eff8db709d93735fa"} Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.299488 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mlwzz" event={"ID":"32d7ebdb-b4ae-4d70-ace7-458c7334b352","Type":"ContainerStarted","Data":"8887a423d048c2113f2c63110a21f2dd8829decdd51bb095be8c860134624f17"} Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.299540 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.299568 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.303095 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gl88h" event={"ID":"9cc73ea6-7c30-4a29-be40-7b85ceaeff71","Type":"ContainerStarted","Data":"0016c7d8733f836bc0a5af1a59d8ca3fbf7d8c557e229042dae01437f2913aa6"} Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.327672 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-mlwzz" podStartSLOduration=4.327655331 podStartE2EDuration="4.327655331s" podCreationTimestamp="2025-09-29 20:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:45:58.314215675 +0000 UTC m=+5799.962005017" watchObservedRunningTime="2025-09-29 20:45:58.327655331 +0000 UTC m=+5799.975444663" Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.340413 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-gl88h" podStartSLOduration=2.340397426 podStartE2EDuration="2.340397426s" podCreationTimestamp="2025-09-29 20:45:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:45:58.336256627 +0000 UTC m=+5799.984045959" watchObservedRunningTime="2025-09-29 20:45:58.340397426 +0000 UTC m=+5799.988186748" Sep 29 20:45:58 crc kubenswrapper[4741]: I0929 20:45:58.450338 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-fsrvf"] Sep 29 20:45:58 crc kubenswrapper[4741]: W0929 20:45:58.450653 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeda0fe8_0c19_4799_a2de_45c6a10014cb.slice/crio-a8537babc86a32afcd1687c85d774e4f4586a7cba5c6a9060bd9ee8fb6d08c27 WatchSource:0}: Error finding container a8537babc86a32afcd1687c85d774e4f4586a7cba5c6a9060bd9ee8fb6d08c27: Status 404 returned error can't find the container with id a8537babc86a32afcd1687c85d774e4f4586a7cba5c6a9060bd9ee8fb6d08c27 Sep 29 20:45:59 crc kubenswrapper[4741]: I0929 20:45:59.310850 4741 generic.go:334] "Generic (PLEG): container finished" podID="deda0fe8-0c19-4799-a2de-45c6a10014cb" containerID="700a401ff67817edf51f2c71f55522cf1c31134d6d9bce8963cebd0340e9955a" exitCode=0 Sep 29 20:45:59 crc kubenswrapper[4741]: I0929 20:45:59.312701 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fsrvf" event={"ID":"deda0fe8-0c19-4799-a2de-45c6a10014cb","Type":"ContainerDied","Data":"700a401ff67817edf51f2c71f55522cf1c31134d6d9bce8963cebd0340e9955a"} Sep 29 20:45:59 crc kubenswrapper[4741]: I0929 20:45:59.312761 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fsrvf" event={"ID":"deda0fe8-0c19-4799-a2de-45c6a10014cb","Type":"ContainerStarted","Data":"a8537babc86a32afcd1687c85d774e4f4586a7cba5c6a9060bd9ee8fb6d08c27"} Sep 29 20:46:00 crc kubenswrapper[4741]: I0929 20:46:00.669913 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fsrvf" Sep 29 20:46:00 crc kubenswrapper[4741]: I0929 20:46:00.716562 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glc8z\" (UniqueName: \"kubernetes.io/projected/deda0fe8-0c19-4799-a2de-45c6a10014cb-kube-api-access-glc8z\") pod \"deda0fe8-0c19-4799-a2de-45c6a10014cb\" (UID: \"deda0fe8-0c19-4799-a2de-45c6a10014cb\") " Sep 29 20:46:00 crc kubenswrapper[4741]: I0929 20:46:00.721098 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deda0fe8-0c19-4799-a2de-45c6a10014cb-kube-api-access-glc8z" (OuterVolumeSpecName: "kube-api-access-glc8z") pod "deda0fe8-0c19-4799-a2de-45c6a10014cb" (UID: "deda0fe8-0c19-4799-a2de-45c6a10014cb"). InnerVolumeSpecName "kube-api-access-glc8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:46:00 crc kubenswrapper[4741]: I0929 20:46:00.818259 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glc8z\" (UniqueName: \"kubernetes.io/projected/deda0fe8-0c19-4799-a2de-45c6a10014cb-kube-api-access-glc8z\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:01 crc kubenswrapper[4741]: I0929 20:46:01.334695 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fsrvf" event={"ID":"deda0fe8-0c19-4799-a2de-45c6a10014cb","Type":"ContainerDied","Data":"a8537babc86a32afcd1687c85d774e4f4586a7cba5c6a9060bd9ee8fb6d08c27"} Sep 29 20:46:01 crc kubenswrapper[4741]: I0929 20:46:01.334741 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8537babc86a32afcd1687c85d774e4f4586a7cba5c6a9060bd9ee8fb6d08c27" Sep 29 20:46:01 crc kubenswrapper[4741]: I0929 20:46:01.334745 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fsrvf" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.643077 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-9899-account-create-5vxm7"] Sep 29 20:46:09 crc kubenswrapper[4741]: E0929 20:46:09.643919 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deda0fe8-0c19-4799-a2de-45c6a10014cb" containerName="mariadb-database-create" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.643937 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="deda0fe8-0c19-4799-a2de-45c6a10014cb" containerName="mariadb-database-create" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.644120 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="deda0fe8-0c19-4799-a2de-45c6a10014cb" containerName="mariadb-database-create" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.644714 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9899-account-create-5vxm7" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.646831 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.652244 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-9899-account-create-5vxm7"] Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.684595 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2bb9\" (UniqueName: \"kubernetes.io/projected/3b2fee03-161c-4a66-8abd-0d793e229153-kube-api-access-l2bb9\") pod \"octavia-9899-account-create-5vxm7\" (UID: \"3b2fee03-161c-4a66-8abd-0d793e229153\") " pod="openstack/octavia-9899-account-create-5vxm7" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.786574 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2bb9\" (UniqueName: \"kubernetes.io/projected/3b2fee03-161c-4a66-8abd-0d793e229153-kube-api-access-l2bb9\") pod \"octavia-9899-account-create-5vxm7\" (UID: \"3b2fee03-161c-4a66-8abd-0d793e229153\") " pod="openstack/octavia-9899-account-create-5vxm7" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.804843 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2bb9\" (UniqueName: \"kubernetes.io/projected/3b2fee03-161c-4a66-8abd-0d793e229153-kube-api-access-l2bb9\") pod \"octavia-9899-account-create-5vxm7\" (UID: \"3b2fee03-161c-4a66-8abd-0d793e229153\") " pod="openstack/octavia-9899-account-create-5vxm7" Sep 29 20:46:09 crc kubenswrapper[4741]: I0929 20:46:09.969092 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9899-account-create-5vxm7" Sep 29 20:46:10 crc kubenswrapper[4741]: I0929 20:46:10.404205 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-9899-account-create-5vxm7"] Sep 29 20:46:10 crc kubenswrapper[4741]: I0929 20:46:10.414238 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-9899-account-create-5vxm7" event={"ID":"3b2fee03-161c-4a66-8abd-0d793e229153","Type":"ContainerStarted","Data":"e4969a367381fb73b5fe4df78ec27cb4f113e22019e8c8bdf7865fb4cf9452b4"} Sep 29 20:46:11 crc kubenswrapper[4741]: I0929 20:46:11.425099 4741 generic.go:334] "Generic (PLEG): container finished" podID="3b2fee03-161c-4a66-8abd-0d793e229153" containerID="a95b5600cb9cc4cf69f11a889eff3455aaa0b97c8a992ebec97d924cae316651" exitCode=0 Sep 29 20:46:11 crc kubenswrapper[4741]: I0929 20:46:11.425228 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-9899-account-create-5vxm7" event={"ID":"3b2fee03-161c-4a66-8abd-0d793e229153","Type":"ContainerDied","Data":"a95b5600cb9cc4cf69f11a889eff3455aaa0b97c8a992ebec97d924cae316651"} Sep 29 20:46:12 crc kubenswrapper[4741]: I0929 20:46:12.879408 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9899-account-create-5vxm7" Sep 29 20:46:12 crc kubenswrapper[4741]: I0929 20:46:12.975078 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2bb9\" (UniqueName: \"kubernetes.io/projected/3b2fee03-161c-4a66-8abd-0d793e229153-kube-api-access-l2bb9\") pod \"3b2fee03-161c-4a66-8abd-0d793e229153\" (UID: \"3b2fee03-161c-4a66-8abd-0d793e229153\") " Sep 29 20:46:12 crc kubenswrapper[4741]: I0929 20:46:12.981305 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b2fee03-161c-4a66-8abd-0d793e229153-kube-api-access-l2bb9" (OuterVolumeSpecName: "kube-api-access-l2bb9") pod "3b2fee03-161c-4a66-8abd-0d793e229153" (UID: "3b2fee03-161c-4a66-8abd-0d793e229153"). InnerVolumeSpecName "kube-api-access-l2bb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:46:13 crc kubenswrapper[4741]: I0929 20:46:13.076864 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2bb9\" (UniqueName: \"kubernetes.io/projected/3b2fee03-161c-4a66-8abd-0d793e229153-kube-api-access-l2bb9\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:13 crc kubenswrapper[4741]: I0929 20:46:13.456519 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-9899-account-create-5vxm7" event={"ID":"3b2fee03-161c-4a66-8abd-0d793e229153","Type":"ContainerDied","Data":"e4969a367381fb73b5fe4df78ec27cb4f113e22019e8c8bdf7865fb4cf9452b4"} Sep 29 20:46:13 crc kubenswrapper[4741]: I0929 20:46:13.456562 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4969a367381fb73b5fe4df78ec27cb4f113e22019e8c8bdf7865fb4cf9452b4" Sep 29 20:46:13 crc kubenswrapper[4741]: I0929 20:46:13.456634 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9899-account-create-5vxm7" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.107990 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-rskwp"] Sep 29 20:46:16 crc kubenswrapper[4741]: E0929 20:46:16.109024 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b2fee03-161c-4a66-8abd-0d793e229153" containerName="mariadb-account-create" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.109042 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b2fee03-161c-4a66-8abd-0d793e229153" containerName="mariadb-account-create" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.109291 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b2fee03-161c-4a66-8abd-0d793e229153" containerName="mariadb-account-create" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.110174 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-rskwp" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.125167 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-rskwp"] Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.142404 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk8wq\" (UniqueName: \"kubernetes.io/projected/e2e774cc-7135-4580-9f0d-82ead3f51c0b-kube-api-access-kk8wq\") pod \"octavia-persistence-db-create-rskwp\" (UID: \"e2e774cc-7135-4580-9f0d-82ead3f51c0b\") " pod="openstack/octavia-persistence-db-create-rskwp" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.244887 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk8wq\" (UniqueName: \"kubernetes.io/projected/e2e774cc-7135-4580-9f0d-82ead3f51c0b-kube-api-access-kk8wq\") pod \"octavia-persistence-db-create-rskwp\" (UID: \"e2e774cc-7135-4580-9f0d-82ead3f51c0b\") " pod="openstack/octavia-persistence-db-create-rskwp" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.283449 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk8wq\" (UniqueName: \"kubernetes.io/projected/e2e774cc-7135-4580-9f0d-82ead3f51c0b-kube-api-access-kk8wq\") pod \"octavia-persistence-db-create-rskwp\" (UID: \"e2e774cc-7135-4580-9f0d-82ead3f51c0b\") " pod="openstack/octavia-persistence-db-create-rskwp" Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.430262 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-rskwp" Sep 29 20:46:16 crc kubenswrapper[4741]: W0929 20:46:16.886861 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2e774cc_7135_4580_9f0d_82ead3f51c0b.slice/crio-e5141f0c0ad68d45c7219e41ac9be39ff444844545955e0500a97b0e2c188360 WatchSource:0}: Error finding container e5141f0c0ad68d45c7219e41ac9be39ff444844545955e0500a97b0e2c188360: Status 404 returned error can't find the container with id e5141f0c0ad68d45c7219e41ac9be39ff444844545955e0500a97b0e2c188360 Sep 29 20:46:16 crc kubenswrapper[4741]: I0929 20:46:16.887816 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-rskwp"] Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.491761 4741 generic.go:334] "Generic (PLEG): container finished" podID="e2e774cc-7135-4580-9f0d-82ead3f51c0b" containerID="12e42207162c5b9dae00c1f4d62d733c8d650ab7de27d525d7419ffb9c7a2859" exitCode=0 Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.491807 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-rskwp" event={"ID":"e2e774cc-7135-4580-9f0d-82ead3f51c0b","Type":"ContainerDied","Data":"12e42207162c5b9dae00c1f4d62d733c8d650ab7de27d525d7419ffb9c7a2859"} Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.491840 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-rskwp" event={"ID":"e2e774cc-7135-4580-9f0d-82ead3f51c0b","Type":"ContainerStarted","Data":"e5141f0c0ad68d45c7219e41ac9be39ff444844545955e0500a97b0e2c188360"} Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.730377 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-95pqk"] Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.732731 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.744903 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-95pqk"] Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.779447 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-catalog-content\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.779681 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-utilities\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.779741 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76zh6\" (UniqueName: \"kubernetes.io/projected/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-kube-api-access-76zh6\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.881717 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76zh6\" (UniqueName: \"kubernetes.io/projected/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-kube-api-access-76zh6\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.882078 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-catalog-content\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.882199 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-utilities\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.882762 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-catalog-content\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.882779 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-utilities\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:17 crc kubenswrapper[4741]: I0929 20:46:17.908474 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76zh6\" (UniqueName: \"kubernetes.io/projected/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-kube-api-access-76zh6\") pod \"redhat-operators-95pqk\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:18 crc kubenswrapper[4741]: I0929 20:46:18.051153 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:18 crc kubenswrapper[4741]: I0929 20:46:18.329870 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-95pqk"] Sep 29 20:46:18 crc kubenswrapper[4741]: W0929 20:46:18.341052 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeea6ae27_6f3c_40d0_9ffe_1644ab965ef6.slice/crio-d4a1cadc8980d9c4a2f5cc17dcd81d478ce4d13412fb7f8b10d7a1442b931130 WatchSource:0}: Error finding container d4a1cadc8980d9c4a2f5cc17dcd81d478ce4d13412fb7f8b10d7a1442b931130: Status 404 returned error can't find the container with id d4a1cadc8980d9c4a2f5cc17dcd81d478ce4d13412fb7f8b10d7a1442b931130 Sep 29 20:46:18 crc kubenswrapper[4741]: I0929 20:46:18.501992 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95pqk" event={"ID":"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6","Type":"ContainerStarted","Data":"d4a1cadc8980d9c4a2f5cc17dcd81d478ce4d13412fb7f8b10d7a1442b931130"} Sep 29 20:46:18 crc kubenswrapper[4741]: I0929 20:46:18.795658 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-rskwp" Sep 29 20:46:18 crc kubenswrapper[4741]: I0929 20:46:18.901662 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk8wq\" (UniqueName: \"kubernetes.io/projected/e2e774cc-7135-4580-9f0d-82ead3f51c0b-kube-api-access-kk8wq\") pod \"e2e774cc-7135-4580-9f0d-82ead3f51c0b\" (UID: \"e2e774cc-7135-4580-9f0d-82ead3f51c0b\") " Sep 29 20:46:18 crc kubenswrapper[4741]: I0929 20:46:18.908603 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2e774cc-7135-4580-9f0d-82ead3f51c0b-kube-api-access-kk8wq" (OuterVolumeSpecName: "kube-api-access-kk8wq") pod "e2e774cc-7135-4580-9f0d-82ead3f51c0b" (UID: "e2e774cc-7135-4580-9f0d-82ead3f51c0b"). InnerVolumeSpecName "kube-api-access-kk8wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:46:19 crc kubenswrapper[4741]: I0929 20:46:19.004503 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk8wq\" (UniqueName: \"kubernetes.io/projected/e2e774cc-7135-4580-9f0d-82ead3f51c0b-kube-api-access-kk8wq\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:19 crc kubenswrapper[4741]: I0929 20:46:19.513267 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-rskwp" Sep 29 20:46:19 crc kubenswrapper[4741]: I0929 20:46:19.513212 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-rskwp" event={"ID":"e2e774cc-7135-4580-9f0d-82ead3f51c0b","Type":"ContainerDied","Data":"e5141f0c0ad68d45c7219e41ac9be39ff444844545955e0500a97b0e2c188360"} Sep 29 20:46:19 crc kubenswrapper[4741]: I0929 20:46:19.513600 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5141f0c0ad68d45c7219e41ac9be39ff444844545955e0500a97b0e2c188360" Sep 29 20:46:19 crc kubenswrapper[4741]: I0929 20:46:19.515612 4741 generic.go:334] "Generic (PLEG): container finished" podID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerID="1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e" exitCode=0 Sep 29 20:46:19 crc kubenswrapper[4741]: I0929 20:46:19.515643 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95pqk" event={"ID":"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6","Type":"ContainerDied","Data":"1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e"} Sep 29 20:46:20 crc kubenswrapper[4741]: I0929 20:46:20.527775 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95pqk" event={"ID":"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6","Type":"ContainerStarted","Data":"4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d"} Sep 29 20:46:22 crc kubenswrapper[4741]: I0929 20:46:22.551897 4741 generic.go:334] "Generic (PLEG): container finished" podID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerID="4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d" exitCode=0 Sep 29 20:46:22 crc kubenswrapper[4741]: I0929 20:46:22.552001 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95pqk" event={"ID":"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6","Type":"ContainerDied","Data":"4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d"} Sep 29 20:46:23 crc kubenswrapper[4741]: I0929 20:46:23.569089 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95pqk" event={"ID":"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6","Type":"ContainerStarted","Data":"0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b"} Sep 29 20:46:23 crc kubenswrapper[4741]: I0929 20:46:23.599347 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-95pqk" podStartSLOduration=3.135728074 podStartE2EDuration="6.599326878s" podCreationTimestamp="2025-09-29 20:46:17 +0000 UTC" firstStartedPulling="2025-09-29 20:46:19.517704164 +0000 UTC m=+5821.165493496" lastFinishedPulling="2025-09-29 20:46:22.981302968 +0000 UTC m=+5824.629092300" observedRunningTime="2025-09-29 20:46:23.593971535 +0000 UTC m=+5825.241760877" watchObservedRunningTime="2025-09-29 20:46:23.599326878 +0000 UTC m=+5825.247116210" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.171877 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-f535-account-create-drz2b"] Sep 29 20:46:27 crc kubenswrapper[4741]: E0929 20:46:27.172584 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e774cc-7135-4580-9f0d-82ead3f51c0b" containerName="mariadb-database-create" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.172598 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e774cc-7135-4580-9f0d-82ead3f51c0b" containerName="mariadb-database-create" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.172801 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2e774cc-7135-4580-9f0d-82ead3f51c0b" containerName="mariadb-database-create" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.173468 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f535-account-create-drz2b" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.175577 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.198703 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-f535-account-create-drz2b"] Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.281794 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffdvw\" (UniqueName: \"kubernetes.io/projected/cb715a13-36f6-4750-baa3-1f51eeb3856d-kube-api-access-ffdvw\") pod \"octavia-f535-account-create-drz2b\" (UID: \"cb715a13-36f6-4750-baa3-1f51eeb3856d\") " pod="openstack/octavia-f535-account-create-drz2b" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.383261 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffdvw\" (UniqueName: \"kubernetes.io/projected/cb715a13-36f6-4750-baa3-1f51eeb3856d-kube-api-access-ffdvw\") pod \"octavia-f535-account-create-drz2b\" (UID: \"cb715a13-36f6-4750-baa3-1f51eeb3856d\") " pod="openstack/octavia-f535-account-create-drz2b" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.403278 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffdvw\" (UniqueName: \"kubernetes.io/projected/cb715a13-36f6-4750-baa3-1f51eeb3856d-kube-api-access-ffdvw\") pod \"octavia-f535-account-create-drz2b\" (UID: \"cb715a13-36f6-4750-baa3-1f51eeb3856d\") " pod="openstack/octavia-f535-account-create-drz2b" Sep 29 20:46:27 crc kubenswrapper[4741]: I0929 20:46:27.502138 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f535-account-create-drz2b" Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.011027 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-f535-account-create-drz2b"] Sep 29 20:46:28 crc kubenswrapper[4741]: W0929 20:46:28.014721 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb715a13_36f6_4750_baa3_1f51eeb3856d.slice/crio-cb486b05fa0eb1559c700426e47ccfba64cd63917bef411bf33e9aab80ff583b WatchSource:0}: Error finding container cb486b05fa0eb1559c700426e47ccfba64cd63917bef411bf33e9aab80ff583b: Status 404 returned error can't find the container with id cb486b05fa0eb1559c700426e47ccfba64cd63917bef411bf33e9aab80ff583b Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.052173 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.053368 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.125004 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.622009 4741 generic.go:334] "Generic (PLEG): container finished" podID="cb715a13-36f6-4750-baa3-1f51eeb3856d" containerID="064b2dbb039ec7f838bf18a3a72321c366b352a397bacad477ea937782e2c581" exitCode=0 Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.622205 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f535-account-create-drz2b" event={"ID":"cb715a13-36f6-4750-baa3-1f51eeb3856d","Type":"ContainerDied","Data":"064b2dbb039ec7f838bf18a3a72321c366b352a397bacad477ea937782e2c581"} Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.623923 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f535-account-create-drz2b" event={"ID":"cb715a13-36f6-4750-baa3-1f51eeb3856d","Type":"ContainerStarted","Data":"cb486b05fa0eb1559c700426e47ccfba64cd63917bef411bf33e9aab80ff583b"} Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.690920 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:28 crc kubenswrapper[4741]: I0929 20:46:28.743765 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-95pqk"] Sep 29 20:46:29 crc kubenswrapper[4741]: I0929 20:46:29.985567 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f535-account-create-drz2b" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.032488 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffdvw\" (UniqueName: \"kubernetes.io/projected/cb715a13-36f6-4750-baa3-1f51eeb3856d-kube-api-access-ffdvw\") pod \"cb715a13-36f6-4750-baa3-1f51eeb3856d\" (UID: \"cb715a13-36f6-4750-baa3-1f51eeb3856d\") " Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.039349 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb715a13-36f6-4750-baa3-1f51eeb3856d-kube-api-access-ffdvw" (OuterVolumeSpecName: "kube-api-access-ffdvw") pod "cb715a13-36f6-4750-baa3-1f51eeb3856d" (UID: "cb715a13-36f6-4750-baa3-1f51eeb3856d"). InnerVolumeSpecName "kube-api-access-ffdvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.134423 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffdvw\" (UniqueName: \"kubernetes.io/projected/cb715a13-36f6-4750-baa3-1f51eeb3856d-kube-api-access-ffdvw\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.228536 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xbqvm" podUID="8755cb82-1296-4625-8767-4033013502ca" containerName="ovn-controller" probeResult="failure" output=< Sep 29 20:46:30 crc kubenswrapper[4741]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 29 20:46:30 crc kubenswrapper[4741]: > Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.272117 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.274969 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mlwzz" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.380413 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xbqvm-config-dk2w9"] Sep 29 20:46:30 crc kubenswrapper[4741]: E0929 20:46:30.380811 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb715a13-36f6-4750-baa3-1f51eeb3856d" containerName="mariadb-account-create" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.380828 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb715a13-36f6-4750-baa3-1f51eeb3856d" containerName="mariadb-account-create" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.381035 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb715a13-36f6-4750-baa3-1f51eeb3856d" containerName="mariadb-account-create" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.381633 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.384920 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.392287 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xbqvm-config-dk2w9"] Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.439943 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-log-ovn\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.439993 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run-ovn\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.440187 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqbpm\" (UniqueName: \"kubernetes.io/projected/d856209b-14ba-40e6-803d-d76976435552-kube-api-access-nqbpm\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.440245 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-additional-scripts\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.440281 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-scripts\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.440309 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541517 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqbpm\" (UniqueName: \"kubernetes.io/projected/d856209b-14ba-40e6-803d-d76976435552-kube-api-access-nqbpm\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541582 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-additional-scripts\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541611 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-scripts\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541647 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541690 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-log-ovn\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541709 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run-ovn\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541970 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run-ovn\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.541990 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-log-ovn\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.542020 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.542567 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-additional-scripts\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.543438 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-scripts\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.570829 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqbpm\" (UniqueName: \"kubernetes.io/projected/d856209b-14ba-40e6-803d-d76976435552-kube-api-access-nqbpm\") pod \"ovn-controller-xbqvm-config-dk2w9\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.640991 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f535-account-create-drz2b" event={"ID":"cb715a13-36f6-4750-baa3-1f51eeb3856d","Type":"ContainerDied","Data":"cb486b05fa0eb1559c700426e47ccfba64cd63917bef411bf33e9aab80ff583b"} Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.641040 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb486b05fa0eb1559c700426e47ccfba64cd63917bef411bf33e9aab80ff583b" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.641198 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f535-account-create-drz2b" Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.641538 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-95pqk" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="registry-server" containerID="cri-o://0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b" gracePeriod=2 Sep 29 20:46:30 crc kubenswrapper[4741]: I0929 20:46:30.700362 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.222591 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.334581 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xbqvm-config-dk2w9"] Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.378794 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-utilities\") pod \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.379237 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-catalog-content\") pod \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.379797 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-utilities" (OuterVolumeSpecName: "utilities") pod "eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" (UID: "eea6ae27-6f3c-40d0-9ffe-1644ab965ef6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.390739 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76zh6\" (UniqueName: \"kubernetes.io/projected/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-kube-api-access-76zh6\") pod \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\" (UID: \"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6\") " Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.391790 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.395711 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-kube-api-access-76zh6" (OuterVolumeSpecName: "kube-api-access-76zh6") pod "eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" (UID: "eea6ae27-6f3c-40d0-9ffe-1644ab965ef6"). InnerVolumeSpecName "kube-api-access-76zh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.450746 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" (UID: "eea6ae27-6f3c-40d0-9ffe-1644ab965ef6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.493818 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.493896 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76zh6\" (UniqueName: \"kubernetes.io/projected/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6-kube-api-access-76zh6\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.652593 4741 generic.go:334] "Generic (PLEG): container finished" podID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerID="0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b" exitCode=0 Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.652659 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95pqk" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.652667 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95pqk" event={"ID":"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6","Type":"ContainerDied","Data":"0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b"} Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.652723 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95pqk" event={"ID":"eea6ae27-6f3c-40d0-9ffe-1644ab965ef6","Type":"ContainerDied","Data":"d4a1cadc8980d9c4a2f5cc17dcd81d478ce4d13412fb7f8b10d7a1442b931130"} Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.652757 4741 scope.go:117] "RemoveContainer" containerID="0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.657379 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm-config-dk2w9" event={"ID":"d856209b-14ba-40e6-803d-d76976435552","Type":"ContainerStarted","Data":"4db70c09719aa843b9490e6f59cba797fa57f71c212f4752942fcd5447a1f1e9"} Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.657507 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm-config-dk2w9" event={"ID":"d856209b-14ba-40e6-803d-d76976435552","Type":"ContainerStarted","Data":"e99441827e2b4841e46d8e5949735c2fdbe3ecd31ff59b85602686c395db9df4"} Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.676325 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xbqvm-config-dk2w9" podStartSLOduration=1.676304924 podStartE2EDuration="1.676304924s" podCreationTimestamp="2025-09-29 20:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:46:31.675868329 +0000 UTC m=+5833.323657671" watchObservedRunningTime="2025-09-29 20:46:31.676304924 +0000 UTC m=+5833.324094246" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.689351 4741 scope.go:117] "RemoveContainer" containerID="4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.702049 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-95pqk"] Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.715849 4741 scope.go:117] "RemoveContainer" containerID="1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.719151 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-95pqk"] Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.733579 4741 scope.go:117] "RemoveContainer" containerID="0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b" Sep 29 20:46:31 crc kubenswrapper[4741]: E0929 20:46:31.733909 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b\": container with ID starting with 0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b not found: ID does not exist" containerID="0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.733937 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b"} err="failed to get container status \"0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b\": rpc error: code = NotFound desc = could not find container \"0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b\": container with ID starting with 0f08983cc8b730388fb14b92c39e6e037b97edd6a3f174f0b74a5f1c58db6d2b not found: ID does not exist" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.733965 4741 scope.go:117] "RemoveContainer" containerID="4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d" Sep 29 20:46:31 crc kubenswrapper[4741]: E0929 20:46:31.734289 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d\": container with ID starting with 4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d not found: ID does not exist" containerID="4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.734311 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d"} err="failed to get container status \"4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d\": rpc error: code = NotFound desc = could not find container \"4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d\": container with ID starting with 4ef66a9dcaa622800c4a67890e71c8171dc813af1008321a76ae06ccf3c6923d not found: ID does not exist" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.734328 4741 scope.go:117] "RemoveContainer" containerID="1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e" Sep 29 20:46:31 crc kubenswrapper[4741]: E0929 20:46:31.734847 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e\": container with ID starting with 1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e not found: ID does not exist" containerID="1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e" Sep 29 20:46:31 crc kubenswrapper[4741]: I0929 20:46:31.734889 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e"} err="failed to get container status \"1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e\": rpc error: code = NotFound desc = could not find container \"1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e\": container with ID starting with 1d3b09f61692fc2dbb431fb7a4f87f2e70ca6769f6184e2f5fab0a6bf46d258e not found: ID does not exist" Sep 29 20:46:31 crc kubenswrapper[4741]: E0929 20:46:31.831676 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeea6ae27_6f3c_40d0_9ffe_1644ab965ef6.slice/crio-d4a1cadc8980d9c4a2f5cc17dcd81d478ce4d13412fb7f8b10d7a1442b931130\": RecentStats: unable to find data in memory cache]" Sep 29 20:46:32 crc kubenswrapper[4741]: I0929 20:46:32.669368 4741 generic.go:334] "Generic (PLEG): container finished" podID="d856209b-14ba-40e6-803d-d76976435552" containerID="4db70c09719aa843b9490e6f59cba797fa57f71c212f4752942fcd5447a1f1e9" exitCode=0 Sep 29 20:46:32 crc kubenswrapper[4741]: I0929 20:46:32.670552 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm-config-dk2w9" event={"ID":"d856209b-14ba-40e6-803d-d76976435552","Type":"ContainerDied","Data":"4db70c09719aa843b9490e6f59cba797fa57f71c212f4752942fcd5447a1f1e9"} Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.101802 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" path="/var/lib/kubelet/pods/eea6ae27-6f3c-40d0-9ffe-1644ab965ef6/volumes" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.884539 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-868985f9c4-bh7cz"] Sep 29 20:46:33 crc kubenswrapper[4741]: E0929 20:46:33.884912 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="extract-utilities" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.884923 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="extract-utilities" Sep 29 20:46:33 crc kubenswrapper[4741]: E0929 20:46:33.884941 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="extract-content" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.884948 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="extract-content" Sep 29 20:46:33 crc kubenswrapper[4741]: E0929 20:46:33.884959 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="registry-server" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.884965 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="registry-server" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.885214 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="eea6ae27-6f3c-40d0-9ffe-1644ab965ef6" containerName="registry-server" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.891538 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.895116 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.895386 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.895546 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-t99hj" Sep 29 20:46:33 crc kubenswrapper[4741]: I0929 20:46:33.909484 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-868985f9c4-bh7cz"] Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.042768 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-scripts\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.042900 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-config-data\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.042971 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-octavia-run\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.042989 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-combined-ca-bundle\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.043028 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-config-data-merged\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.061100 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.144545 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-log-ovn\") pod \"d856209b-14ba-40e6-803d-d76976435552\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.144624 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqbpm\" (UniqueName: \"kubernetes.io/projected/d856209b-14ba-40e6-803d-d76976435552-kube-api-access-nqbpm\") pod \"d856209b-14ba-40e6-803d-d76976435552\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.144747 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run\") pod \"d856209b-14ba-40e6-803d-d76976435552\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.144775 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run-ovn\") pod \"d856209b-14ba-40e6-803d-d76976435552\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.144881 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-scripts\") pod \"d856209b-14ba-40e6-803d-d76976435552\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145055 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-additional-scripts\") pod \"d856209b-14ba-40e6-803d-d76976435552\" (UID: \"d856209b-14ba-40e6-803d-d76976435552\") " Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145128 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run" (OuterVolumeSpecName: "var-run") pod "d856209b-14ba-40e6-803d-d76976435552" (UID: "d856209b-14ba-40e6-803d-d76976435552"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145176 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d856209b-14ba-40e6-803d-d76976435552" (UID: "d856209b-14ba-40e6-803d-d76976435552"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145405 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-scripts\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145545 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-config-data\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145708 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-octavia-run\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145739 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-combined-ca-bundle\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145859 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-config-data-merged\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145933 4741 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.145953 4741 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.146483 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-config-data-merged\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.146539 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d856209b-14ba-40e6-803d-d76976435552" (UID: "d856209b-14ba-40e6-803d-d76976435552"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.147731 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-octavia-run\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.147874 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d856209b-14ba-40e6-803d-d76976435552" (UID: "d856209b-14ba-40e6-803d-d76976435552"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.148618 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-scripts" (OuterVolumeSpecName: "scripts") pod "d856209b-14ba-40e6-803d-d76976435552" (UID: "d856209b-14ba-40e6-803d-d76976435552"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.151546 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-config-data\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.152108 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d856209b-14ba-40e6-803d-d76976435552-kube-api-access-nqbpm" (OuterVolumeSpecName: "kube-api-access-nqbpm") pod "d856209b-14ba-40e6-803d-d76976435552" (UID: "d856209b-14ba-40e6-803d-d76976435552"). InnerVolumeSpecName "kube-api-access-nqbpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.159614 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-scripts\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.160745 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d39261b-e2bb-4ff2-8e37-1e91d0161c3e-combined-ca-bundle\") pod \"octavia-api-868985f9c4-bh7cz\" (UID: \"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e\") " pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.232405 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.247938 4741 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d856209b-14ba-40e6-803d-d76976435552-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.247977 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.247985 4741 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d856209b-14ba-40e6-803d-d76976435552-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.247996 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqbpm\" (UniqueName: \"kubernetes.io/projected/d856209b-14ba-40e6-803d-d76976435552-kube-api-access-nqbpm\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.687347 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm-config-dk2w9" event={"ID":"d856209b-14ba-40e6-803d-d76976435552","Type":"ContainerDied","Data":"e99441827e2b4841e46d8e5949735c2fdbe3ecd31ff59b85602686c395db9df4"} Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.687713 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e99441827e2b4841e46d8e5949735c2fdbe3ecd31ff59b85602686c395db9df4" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.687422 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-dk2w9" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.751384 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-868985f9c4-bh7cz"] Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.770780 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xbqvm-config-dk2w9"] Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.780299 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xbqvm-config-dk2w9"] Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.891921 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xbqvm-config-j9vnr"] Sep 29 20:46:34 crc kubenswrapper[4741]: E0929 20:46:34.892609 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856209b-14ba-40e6-803d-d76976435552" containerName="ovn-config" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.892624 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856209b-14ba-40e6-803d-d76976435552" containerName="ovn-config" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.893707 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856209b-14ba-40e6-803d-d76976435552" containerName="ovn-config" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.894977 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.899402 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.906795 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xbqvm-config-j9vnr"] Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.970175 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-additional-scripts\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.970274 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-log-ovn\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.970307 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-scripts\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.970335 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.970359 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzrh8\" (UniqueName: \"kubernetes.io/projected/2b6029a1-f709-4fa8-9f05-5377e25beba2-kube-api-access-nzrh8\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:34 crc kubenswrapper[4741]: I0929 20:46:34.970376 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run-ovn\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.071920 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-additional-scripts\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.071992 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-log-ovn\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.072033 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-scripts\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.072061 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.072084 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzrh8\" (UniqueName: \"kubernetes.io/projected/2b6029a1-f709-4fa8-9f05-5377e25beba2-kube-api-access-nzrh8\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.072103 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run-ovn\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.072350 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run-ovn\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.073008 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-additional-scripts\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.073064 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.073442 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-log-ovn\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.074798 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-scripts\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.097725 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzrh8\" (UniqueName: \"kubernetes.io/projected/2b6029a1-f709-4fa8-9f05-5377e25beba2-kube-api-access-nzrh8\") pod \"ovn-controller-xbqvm-config-j9vnr\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.099704 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d856209b-14ba-40e6-803d-d76976435552" path="/var/lib/kubelet/pods/d856209b-14ba-40e6-803d-d76976435552/volumes" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.232163 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-xbqvm" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.233805 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.706908 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-868985f9c4-bh7cz" event={"ID":"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e","Type":"ContainerStarted","Data":"4e2f268d6bd6da9da3723a34e32e350e1b61b532c90350cbe3e0f51ae6debde4"} Sep 29 20:46:35 crc kubenswrapper[4741]: I0929 20:46:35.711859 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xbqvm-config-j9vnr"] Sep 29 20:46:36 crc kubenswrapper[4741]: I0929 20:46:36.719370 4741 generic.go:334] "Generic (PLEG): container finished" podID="2b6029a1-f709-4fa8-9f05-5377e25beba2" containerID="841798414a0df3202bdd06741118227e51058c7cf49d891fd766cdd9ee9464f1" exitCode=0 Sep 29 20:46:36 crc kubenswrapper[4741]: I0929 20:46:36.719533 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm-config-j9vnr" event={"ID":"2b6029a1-f709-4fa8-9f05-5377e25beba2","Type":"ContainerDied","Data":"841798414a0df3202bdd06741118227e51058c7cf49d891fd766cdd9ee9464f1"} Sep 29 20:46:36 crc kubenswrapper[4741]: I0929 20:46:36.719930 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm-config-j9vnr" event={"ID":"2b6029a1-f709-4fa8-9f05-5377e25beba2","Type":"ContainerStarted","Data":"d1c4e1bef3e3e91cafed52c6a77fa7bdf79e1aeda589a3e527f0518e78d0aaf7"} Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.358024 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.449495 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run-ovn\") pod \"2b6029a1-f709-4fa8-9f05-5377e25beba2\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.449644 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzrh8\" (UniqueName: \"kubernetes.io/projected/2b6029a1-f709-4fa8-9f05-5377e25beba2-kube-api-access-nzrh8\") pod \"2b6029a1-f709-4fa8-9f05-5377e25beba2\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.449771 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-log-ovn\") pod \"2b6029a1-f709-4fa8-9f05-5377e25beba2\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.449889 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-scripts\") pod \"2b6029a1-f709-4fa8-9f05-5377e25beba2\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.449930 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-additional-scripts\") pod \"2b6029a1-f709-4fa8-9f05-5377e25beba2\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.449965 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run\") pod \"2b6029a1-f709-4fa8-9f05-5377e25beba2\" (UID: \"2b6029a1-f709-4fa8-9f05-5377e25beba2\") " Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.450469 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "2b6029a1-f709-4fa8-9f05-5377e25beba2" (UID: "2b6029a1-f709-4fa8-9f05-5377e25beba2"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.450502 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "2b6029a1-f709-4fa8-9f05-5377e25beba2" (UID: "2b6029a1-f709-4fa8-9f05-5377e25beba2"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.450540 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run" (OuterVolumeSpecName: "var-run") pod "2b6029a1-f709-4fa8-9f05-5377e25beba2" (UID: "2b6029a1-f709-4fa8-9f05-5377e25beba2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.450895 4741 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.450918 4741 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.450932 4741 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b6029a1-f709-4fa8-9f05-5377e25beba2-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.451059 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "2b6029a1-f709-4fa8-9f05-5377e25beba2" (UID: "2b6029a1-f709-4fa8-9f05-5377e25beba2"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.451367 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-scripts" (OuterVolumeSpecName: "scripts") pod "2b6029a1-f709-4fa8-9f05-5377e25beba2" (UID: "2b6029a1-f709-4fa8-9f05-5377e25beba2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.457712 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b6029a1-f709-4fa8-9f05-5377e25beba2-kube-api-access-nzrh8" (OuterVolumeSpecName: "kube-api-access-nzrh8") pod "2b6029a1-f709-4fa8-9f05-5377e25beba2" (UID: "2b6029a1-f709-4fa8-9f05-5377e25beba2"). InnerVolumeSpecName "kube-api-access-nzrh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.553961 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzrh8\" (UniqueName: \"kubernetes.io/projected/2b6029a1-f709-4fa8-9f05-5377e25beba2-kube-api-access-nzrh8\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.554071 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.554102 4741 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2b6029a1-f709-4fa8-9f05-5377e25beba2-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.794185 4741 generic.go:334] "Generic (PLEG): container finished" podID="5d39261b-e2bb-4ff2-8e37-1e91d0161c3e" containerID="98aa00d8b6739a35eec979b6a45d61a790c0c980d6852280fe48a893da61cd22" exitCode=0 Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.794260 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-868985f9c4-bh7cz" event={"ID":"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e","Type":"ContainerDied","Data":"98aa00d8b6739a35eec979b6a45d61a790c0c980d6852280fe48a893da61cd22"} Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.811524 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xbqvm-config-j9vnr" event={"ID":"2b6029a1-f709-4fa8-9f05-5377e25beba2","Type":"ContainerDied","Data":"d1c4e1bef3e3e91cafed52c6a77fa7bdf79e1aeda589a3e527f0518e78d0aaf7"} Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.811583 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1c4e1bef3e3e91cafed52c6a77fa7bdf79e1aeda589a3e527f0518e78d0aaf7" Sep 29 20:46:43 crc kubenswrapper[4741]: I0929 20:46:43.811560 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xbqvm-config-j9vnr" Sep 29 20:46:44 crc kubenswrapper[4741]: I0929 20:46:44.466853 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xbqvm-config-j9vnr"] Sep 29 20:46:44 crc kubenswrapper[4741]: I0929 20:46:44.477905 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xbqvm-config-j9vnr"] Sep 29 20:46:44 crc kubenswrapper[4741]: I0929 20:46:44.824004 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-868985f9c4-bh7cz" event={"ID":"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e","Type":"ContainerStarted","Data":"7edc5d4a118a493e56612c4bf64fa8accd7c04751cc28d1e5d6036028b60f2bd"} Sep 29 20:46:44 crc kubenswrapper[4741]: I0929 20:46:44.824053 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-868985f9c4-bh7cz" event={"ID":"5d39261b-e2bb-4ff2-8e37-1e91d0161c3e","Type":"ContainerStarted","Data":"e5b10a361f3c1742e99e288bf39eae6f82fe8ea1be79f7a436cadff9e2b2734f"} Sep 29 20:46:44 crc kubenswrapper[4741]: I0929 20:46:44.824159 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:44 crc kubenswrapper[4741]: I0929 20:46:44.824216 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:44 crc kubenswrapper[4741]: I0929 20:46:44.848554 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-868985f9c4-bh7cz" podStartSLOduration=3.293641264 podStartE2EDuration="11.848534287s" podCreationTimestamp="2025-09-29 20:46:33 +0000 UTC" firstStartedPulling="2025-09-29 20:46:34.755534533 +0000 UTC m=+5836.403323865" lastFinishedPulling="2025-09-29 20:46:43.310427536 +0000 UTC m=+5844.958216888" observedRunningTime="2025-09-29 20:46:44.843209346 +0000 UTC m=+5846.490998698" watchObservedRunningTime="2025-09-29 20:46:44.848534287 +0000 UTC m=+5846.496323629" Sep 29 20:46:45 crc kubenswrapper[4741]: I0929 20:46:45.095316 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b6029a1-f709-4fa8-9f05-5377e25beba2" path="/var/lib/kubelet/pods/2b6029a1-f709-4fa8-9f05-5377e25beba2/volumes" Sep 29 20:46:53 crc kubenswrapper[4741]: I0929 20:46:53.122802 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:53 crc kubenswrapper[4741]: I0929 20:46:53.196752 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-868985f9c4-bh7cz" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.794177 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-gjbhp"] Sep 29 20:46:54 crc kubenswrapper[4741]: E0929 20:46:54.795061 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6029a1-f709-4fa8-9f05-5377e25beba2" containerName="ovn-config" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.795079 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6029a1-f709-4fa8-9f05-5377e25beba2" containerName="ovn-config" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.795338 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b6029a1-f709-4fa8-9f05-5377e25beba2" containerName="ovn-config" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.796911 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.800080 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.800286 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.800458 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.812175 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-gjbhp"] Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.969349 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-hm-ports\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.969433 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-config-data-merged\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.969582 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-config-data\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:54 crc kubenswrapper[4741]: I0929 20:46:54.969644 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-scripts\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.071565 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-scripts\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.071653 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-hm-ports\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.071705 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-config-data-merged\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.071889 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-config-data\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.074208 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-config-data-merged\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.074266 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-hm-ports\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.078350 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-scripts\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.112556 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f0963e9-4b39-4a0a-8ba5-7f47226efd00-config-data\") pod \"octavia-rsyslog-gjbhp\" (UID: \"1f0963e9-4b39-4a0a-8ba5-7f47226efd00\") " pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.124727 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.704625 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-gjbhp"] Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.881594 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fwn97"] Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.883588 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.885989 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.893968 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fwn97"] Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.941563 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-gjbhp" event={"ID":"1f0963e9-4b39-4a0a-8ba5-7f47226efd00","Type":"ContainerStarted","Data":"980438d35fb25e388d001d383aad9545bdbecd41a5f3a61faa99cb98f8347b14"} Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.987937 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c8516c2-d162-47bc-b9f4-bca1708e0868-httpd-config\") pod \"octavia-image-upload-59f8cff499-fwn97\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:55 crc kubenswrapper[4741]: I0929 20:46:55.988012 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/3c8516c2-d162-47bc-b9f4-bca1708e0868-amphora-image\") pod \"octavia-image-upload-59f8cff499-fwn97\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:56 crc kubenswrapper[4741]: I0929 20:46:56.090427 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/3c8516c2-d162-47bc-b9f4-bca1708e0868-amphora-image\") pod \"octavia-image-upload-59f8cff499-fwn97\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:56 crc kubenswrapper[4741]: I0929 20:46:56.090655 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c8516c2-d162-47bc-b9f4-bca1708e0868-httpd-config\") pod \"octavia-image-upload-59f8cff499-fwn97\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:56 crc kubenswrapper[4741]: I0929 20:46:56.090851 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/3c8516c2-d162-47bc-b9f4-bca1708e0868-amphora-image\") pod \"octavia-image-upload-59f8cff499-fwn97\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:56 crc kubenswrapper[4741]: I0929 20:46:56.096140 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c8516c2-d162-47bc-b9f4-bca1708e0868-httpd-config\") pod \"octavia-image-upload-59f8cff499-fwn97\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:56 crc kubenswrapper[4741]: I0929 20:46:56.209339 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:46:56 crc kubenswrapper[4741]: I0929 20:46:56.701896 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fwn97"] Sep 29 20:46:57 crc kubenswrapper[4741]: I0929 20:46:57.962924 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-gjbhp" event={"ID":"1f0963e9-4b39-4a0a-8ba5-7f47226efd00","Type":"ContainerStarted","Data":"30aa629bde47bc5203fe09bb0c738a3331bafc107f9ef7f526342d36c9dcc3d0"} Sep 29 20:46:57 crc kubenswrapper[4741]: I0929 20:46:57.964472 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fwn97" event={"ID":"3c8516c2-d162-47bc-b9f4-bca1708e0868","Type":"ContainerStarted","Data":"19962560e7cae1cd51422ef7430faa6d66870637e1397ef1c37ba27f5c26a119"} Sep 29 20:46:59 crc kubenswrapper[4741]: I0929 20:46:59.987081 4741 generic.go:334] "Generic (PLEG): container finished" podID="1f0963e9-4b39-4a0a-8ba5-7f47226efd00" containerID="30aa629bde47bc5203fe09bb0c738a3331bafc107f9ef7f526342d36c9dcc3d0" exitCode=0 Sep 29 20:46:59 crc kubenswrapper[4741]: I0929 20:46:59.987141 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-gjbhp" event={"ID":"1f0963e9-4b39-4a0a-8ba5-7f47226efd00","Type":"ContainerDied","Data":"30aa629bde47bc5203fe09bb0c738a3331bafc107f9ef7f526342d36c9dcc3d0"} Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.271265 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-nsg8t"] Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.273047 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.274835 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.275128 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.275263 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.291096 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-nsg8t"] Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.404250 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-config-data\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.404303 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/13b51e85-b858-4c7d-b92f-d082220fae38-hm-ports\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.404353 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-combined-ca-bundle\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.404427 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-scripts\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.404456 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/13b51e85-b858-4c7d-b92f-d082220fae38-config-data-merged\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.404776 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-amphora-certs\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.506602 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-config-data\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.506679 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/13b51e85-b858-4c7d-b92f-d082220fae38-hm-ports\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.506726 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-combined-ca-bundle\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.506759 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-scripts\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.506786 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/13b51e85-b858-4c7d-b92f-d082220fae38-config-data-merged\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.506882 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-amphora-certs\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.508094 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/13b51e85-b858-4c7d-b92f-d082220fae38-config-data-merged\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.508559 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/13b51e85-b858-4c7d-b92f-d082220fae38-hm-ports\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.517061 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-scripts\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.518008 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-amphora-certs\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.518641 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-combined-ca-bundle\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.522462 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b51e85-b858-4c7d-b92f-d082220fae38-config-data\") pod \"octavia-healthmanager-nsg8t\" (UID: \"13b51e85-b858-4c7d-b92f-d082220fae38\") " pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:01 crc kubenswrapper[4741]: I0929 20:47:01.597254 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.635040 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-rvkpf"] Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.637067 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.646688 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.651611 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-rvkpf"] Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.728033 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-scripts\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.728157 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11060069-6fb7-4371-9be8-938a92c8269e-config-data-merged\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.728207 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-config-data\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.728234 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-combined-ca-bundle\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.829963 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-config-data\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.830020 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-combined-ca-bundle\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.830136 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-scripts\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.830212 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11060069-6fb7-4371-9be8-938a92c8269e-config-data-merged\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.830816 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11060069-6fb7-4371-9be8-938a92c8269e-config-data-merged\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.841203 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-scripts\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.841361 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-combined-ca-bundle\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.842023 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-config-data\") pod \"octavia-db-sync-rvkpf\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:02 crc kubenswrapper[4741]: I0929 20:47:02.965338 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.614981 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-v24jm"] Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.617322 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.629280 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.629348 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.631285 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-v24jm"] Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.730172 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-nsg8t"] Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.768438 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-combined-ca-bundle\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.768539 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-scripts\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.768588 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-config-data\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.768612 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-amphora-certs\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.768922 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-hm-ports\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.769003 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-config-data-merged\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.852823 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-rvkpf"] Sep 29 20:47:04 crc kubenswrapper[4741]: W0929 20:47:04.866723 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11060069_6fb7_4371_9be8_938a92c8269e.slice/crio-2e5f410573ef4c26fdd77e2bd8f6b7299aee34fb7dd3acdff66dabb4cd3cfa12 WatchSource:0}: Error finding container 2e5f410573ef4c26fdd77e2bd8f6b7299aee34fb7dd3acdff66dabb4cd3cfa12: Status 404 returned error can't find the container with id 2e5f410573ef4c26fdd77e2bd8f6b7299aee34fb7dd3acdff66dabb4cd3cfa12 Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.870608 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-config-data\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.870653 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-amphora-certs\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.870733 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-hm-ports\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.870770 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-config-data-merged\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.870838 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-combined-ca-bundle\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.870909 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-scripts\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.872752 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-config-data-merged\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.877637 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-hm-ports\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.877893 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-combined-ca-bundle\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.878008 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-amphora-certs\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.878539 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-config-data\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.882291 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfce8c59-065c-4d36-a85d-ddbb32dc6bf9-scripts\") pod \"octavia-housekeeping-v24jm\" (UID: \"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9\") " pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:04 crc kubenswrapper[4741]: I0929 20:47:04.944040 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.077943 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-gjbhp" event={"ID":"1f0963e9-4b39-4a0a-8ba5-7f47226efd00","Type":"ContainerStarted","Data":"3fe0ec6da1b545f98b3cb3773424fefed6b55e87550f2c7635459fc42d9262ce"} Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.078455 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.080894 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fwn97" event={"ID":"3c8516c2-d162-47bc-b9f4-bca1708e0868","Type":"ContainerStarted","Data":"c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5"} Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.083768 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-nsg8t" event={"ID":"13b51e85-b858-4c7d-b92f-d082220fae38","Type":"ContainerStarted","Data":"e875b1d3b3b53fe78a0df925e9d6f844aad364f18c756ec4b58e40320b188aa2"} Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.085262 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rvkpf" event={"ID":"11060069-6fb7-4371-9be8-938a92c8269e","Type":"ContainerStarted","Data":"0fbd7b96e2257f71ec10e9d0fbaf2b94ca5b1b36b6d1eff6f5afdaaf2bfcaa56"} Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.085292 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rvkpf" event={"ID":"11060069-6fb7-4371-9be8-938a92c8269e","Type":"ContainerStarted","Data":"2e5f410573ef4c26fdd77e2bd8f6b7299aee34fb7dd3acdff66dabb4cd3cfa12"} Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.100460 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-gjbhp" podStartSLOduration=5.690174623 podStartE2EDuration="11.100446999s" podCreationTimestamp="2025-09-29 20:46:54 +0000 UTC" firstStartedPulling="2025-09-29 20:46:55.715382922 +0000 UTC m=+5857.363172254" lastFinishedPulling="2025-09-29 20:47:01.125655298 +0000 UTC m=+5862.773444630" observedRunningTime="2025-09-29 20:47:05.0933353 +0000 UTC m=+5866.741124632" watchObservedRunningTime="2025-09-29 20:47:05.100446999 +0000 UTC m=+5866.748236331" Sep 29 20:47:05 crc kubenswrapper[4741]: I0929 20:47:05.554165 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-v24jm"] Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.096453 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-nsg8t" event={"ID":"13b51e85-b858-4c7d-b92f-d082220fae38","Type":"ContainerStarted","Data":"6dd66d7eb296b82e622c2280c635bb28594dec0a5017257a49ca343552d33b74"} Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.100227 4741 generic.go:334] "Generic (PLEG): container finished" podID="11060069-6fb7-4371-9be8-938a92c8269e" containerID="0fbd7b96e2257f71ec10e9d0fbaf2b94ca5b1b36b6d1eff6f5afdaaf2bfcaa56" exitCode=0 Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.100388 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rvkpf" event={"ID":"11060069-6fb7-4371-9be8-938a92c8269e","Type":"ContainerDied","Data":"0fbd7b96e2257f71ec10e9d0fbaf2b94ca5b1b36b6d1eff6f5afdaaf2bfcaa56"} Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.103500 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-v24jm" event={"ID":"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9","Type":"ContainerStarted","Data":"b5a83bc0459e3b18be6e42c0ce68cac6e324af973a2083e0958a22008f102d4b"} Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.105552 4741 generic.go:334] "Generic (PLEG): container finished" podID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerID="c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5" exitCode=0 Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.105601 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fwn97" event={"ID":"3c8516c2-d162-47bc-b9f4-bca1708e0868","Type":"ContainerDied","Data":"c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5"} Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.287916 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-744s7"] Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.290232 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.293975 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-744s7"] Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.294912 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.296155 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.408291 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-amphora-certs\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.408355 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fba4561b-c4ff-4b6b-8785-e7605776aaf2-config-data-merged\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.408381 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/fba4561b-c4ff-4b6b-8785-e7605776aaf2-hm-ports\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.408479 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-config-data\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.408547 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-combined-ca-bundle\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.408586 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-scripts\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.510295 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-amphora-certs\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.510360 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fba4561b-c4ff-4b6b-8785-e7605776aaf2-config-data-merged\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.510383 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/fba4561b-c4ff-4b6b-8785-e7605776aaf2-hm-ports\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.510424 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-config-data\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.510491 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-combined-ca-bundle\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.510529 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-scripts\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.511607 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fba4561b-c4ff-4b6b-8785-e7605776aaf2-config-data-merged\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.512859 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/fba4561b-c4ff-4b6b-8785-e7605776aaf2-hm-ports\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.515641 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-scripts\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.515957 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-config-data\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.516339 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-combined-ca-bundle\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.531276 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/fba4561b-c4ff-4b6b-8785-e7605776aaf2-amphora-certs\") pod \"octavia-worker-744s7\" (UID: \"fba4561b-c4ff-4b6b-8785-e7605776aaf2\") " pod="openstack/octavia-worker-744s7" Sep 29 20:47:06 crc kubenswrapper[4741]: I0929 20:47:06.633235 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-744s7" Sep 29 20:47:07 crc kubenswrapper[4741]: I0929 20:47:07.139568 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-744s7"] Sep 29 20:47:07 crc kubenswrapper[4741]: W0929 20:47:07.144065 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfba4561b_c4ff_4b6b_8785_e7605776aaf2.slice/crio-a0131a6bad21d01bf8f37e8710955427f5a6d74f33f0d9a2e76f70f4774c0b19 WatchSource:0}: Error finding container a0131a6bad21d01bf8f37e8710955427f5a6d74f33f0d9a2e76f70f4774c0b19: Status 404 returned error can't find the container with id a0131a6bad21d01bf8f37e8710955427f5a6d74f33f0d9a2e76f70f4774c0b19 Sep 29 20:47:08 crc kubenswrapper[4741]: I0929 20:47:08.123900 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-744s7" event={"ID":"fba4561b-c4ff-4b6b-8785-e7605776aaf2","Type":"ContainerStarted","Data":"a0131a6bad21d01bf8f37e8710955427f5a6d74f33f0d9a2e76f70f4774c0b19"} Sep 29 20:47:10 crc kubenswrapper[4741]: I0929 20:47:10.147748 4741 generic.go:334] "Generic (PLEG): container finished" podID="13b51e85-b858-4c7d-b92f-d082220fae38" containerID="6dd66d7eb296b82e622c2280c635bb28594dec0a5017257a49ca343552d33b74" exitCode=0 Sep 29 20:47:10 crc kubenswrapper[4741]: I0929 20:47:10.147835 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-nsg8t" event={"ID":"13b51e85-b858-4c7d-b92f-d082220fae38","Type":"ContainerDied","Data":"6dd66d7eb296b82e622c2280c635bb28594dec0a5017257a49ca343552d33b74"} Sep 29 20:47:10 crc kubenswrapper[4741]: I0929 20:47:10.151613 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rvkpf" event={"ID":"11060069-6fb7-4371-9be8-938a92c8269e","Type":"ContainerStarted","Data":"c1d008ea2aa0a500cbcc52c146bc49b0959bb30e5da2cd55abe927237a3389de"} Sep 29 20:47:10 crc kubenswrapper[4741]: I0929 20:47:10.164658 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-gjbhp" Sep 29 20:47:10 crc kubenswrapper[4741]: I0929 20:47:10.218469 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-rvkpf" podStartSLOduration=8.218445779 podStartE2EDuration="8.218445779s" podCreationTimestamp="2025-09-29 20:47:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:47:10.191124469 +0000 UTC m=+5871.838913811" watchObservedRunningTime="2025-09-29 20:47:10.218445779 +0000 UTC m=+5871.866235111" Sep 29 20:47:11 crc kubenswrapper[4741]: I0929 20:47:11.163350 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-744s7" event={"ID":"fba4561b-c4ff-4b6b-8785-e7605776aaf2","Type":"ContainerStarted","Data":"ae0e7f27a9c2a41ff7623c8c4697940ee4bb44394fe627973c6f90846c8af6a6"} Sep 29 20:47:11 crc kubenswrapper[4741]: I0929 20:47:11.167951 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-v24jm" event={"ID":"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9","Type":"ContainerStarted","Data":"1c76fb57444c5e562e02d519bd7fcda6132de14e7858fdfcc10f54fcae708757"} Sep 29 20:47:11 crc kubenswrapper[4741]: I0929 20:47:11.172044 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fwn97" event={"ID":"3c8516c2-d162-47bc-b9f4-bca1708e0868","Type":"ContainerStarted","Data":"56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770"} Sep 29 20:47:11 crc kubenswrapper[4741]: I0929 20:47:11.180137 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-nsg8t" event={"ID":"13b51e85-b858-4c7d-b92f-d082220fae38","Type":"ContainerStarted","Data":"f3eb668a47a1052a20d7c04041582066dadf2186b2e71653f3ce2a5dfc3ac074"} Sep 29 20:47:11 crc kubenswrapper[4741]: I0929 20:47:11.180616 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:11 crc kubenswrapper[4741]: I0929 20:47:11.217142 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-fwn97" podStartSLOduration=2.484154824 podStartE2EDuration="16.217124182s" podCreationTimestamp="2025-09-29 20:46:55 +0000 UTC" firstStartedPulling="2025-09-29 20:46:56.986972527 +0000 UTC m=+5858.634761859" lastFinishedPulling="2025-09-29 20:47:10.719941885 +0000 UTC m=+5872.367731217" observedRunningTime="2025-09-29 20:47:11.199103502 +0000 UTC m=+5872.846892834" watchObservedRunningTime="2025-09-29 20:47:11.217124182 +0000 UTC m=+5872.864913514" Sep 29 20:47:11 crc kubenswrapper[4741]: I0929 20:47:11.225479 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-nsg8t" podStartSLOduration=10.225462061 podStartE2EDuration="10.225462061s" podCreationTimestamp="2025-09-29 20:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:47:11.220013585 +0000 UTC m=+5872.867802917" watchObservedRunningTime="2025-09-29 20:47:11.225462061 +0000 UTC m=+5872.873251393" Sep 29 20:47:12 crc kubenswrapper[4741]: I0929 20:47:12.195287 4741 generic.go:334] "Generic (PLEG): container finished" podID="dfce8c59-065c-4d36-a85d-ddbb32dc6bf9" containerID="1c76fb57444c5e562e02d519bd7fcda6132de14e7858fdfcc10f54fcae708757" exitCode=0 Sep 29 20:47:12 crc kubenswrapper[4741]: I0929 20:47:12.195606 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-v24jm" event={"ID":"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9","Type":"ContainerDied","Data":"1c76fb57444c5e562e02d519bd7fcda6132de14e7858fdfcc10f54fcae708757"} Sep 29 20:47:12 crc kubenswrapper[4741]: I0929 20:47:12.198090 4741 generic.go:334] "Generic (PLEG): container finished" podID="fba4561b-c4ff-4b6b-8785-e7605776aaf2" containerID="ae0e7f27a9c2a41ff7623c8c4697940ee4bb44394fe627973c6f90846c8af6a6" exitCode=0 Sep 29 20:47:12 crc kubenswrapper[4741]: I0929 20:47:12.198126 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-744s7" event={"ID":"fba4561b-c4ff-4b6b-8785-e7605776aaf2","Type":"ContainerDied","Data":"ae0e7f27a9c2a41ff7623c8c4697940ee4bb44394fe627973c6f90846c8af6a6"} Sep 29 20:47:12 crc kubenswrapper[4741]: I0929 20:47:12.201630 4741 generic.go:334] "Generic (PLEG): container finished" podID="11060069-6fb7-4371-9be8-938a92c8269e" containerID="c1d008ea2aa0a500cbcc52c146bc49b0959bb30e5da2cd55abe927237a3389de" exitCode=0 Sep 29 20:47:12 crc kubenswrapper[4741]: I0929 20:47:12.201729 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rvkpf" event={"ID":"11060069-6fb7-4371-9be8-938a92c8269e","Type":"ContainerDied","Data":"c1d008ea2aa0a500cbcc52c146bc49b0959bb30e5da2cd55abe927237a3389de"} Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.219586 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-v24jm" event={"ID":"dfce8c59-065c-4d36-a85d-ddbb32dc6bf9","Type":"ContainerStarted","Data":"1c87a224f58e9be3b8e01dfa29787739f9a8b27275a33bf5f2d7c34bc41597a9"} Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.220606 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.224194 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-744s7" event={"ID":"fba4561b-c4ff-4b6b-8785-e7605776aaf2","Type":"ContainerStarted","Data":"741bec1c4c1b7ab0ce7953fdbb7243b76dd3d2576ec10fa2d8610d09a47e3bfd"} Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.224411 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-744s7" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.250957 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-v24jm" podStartSLOduration=4.290097095 podStartE2EDuration="9.250938993s" podCreationTimestamp="2025-09-29 20:47:04 +0000 UTC" firstStartedPulling="2025-09-29 20:47:05.568130305 +0000 UTC m=+5867.215919637" lastFinishedPulling="2025-09-29 20:47:10.528972203 +0000 UTC m=+5872.176761535" observedRunningTime="2025-09-29 20:47:13.244954951 +0000 UTC m=+5874.892744293" watchObservedRunningTime="2025-09-29 20:47:13.250938993 +0000 UTC m=+5874.898728315" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.613366 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.637736 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-744s7" podStartSLOduration=4.249082557 podStartE2EDuration="7.637716594s" podCreationTimestamp="2025-09-29 20:47:06 +0000 UTC" firstStartedPulling="2025-09-29 20:47:07.146330099 +0000 UTC m=+5868.794119431" lastFinishedPulling="2025-09-29 20:47:10.534964136 +0000 UTC m=+5872.182753468" observedRunningTime="2025-09-29 20:47:13.265180772 +0000 UTC m=+5874.912970134" watchObservedRunningTime="2025-09-29 20:47:13.637716594 +0000 UTC m=+5875.285505926" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.760237 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-config-data\") pod \"11060069-6fb7-4371-9be8-938a92c8269e\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.760320 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11060069-6fb7-4371-9be8-938a92c8269e-config-data-merged\") pod \"11060069-6fb7-4371-9be8-938a92c8269e\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.760382 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-scripts\") pod \"11060069-6fb7-4371-9be8-938a92c8269e\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.760434 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-combined-ca-bundle\") pod \"11060069-6fb7-4371-9be8-938a92c8269e\" (UID: \"11060069-6fb7-4371-9be8-938a92c8269e\") " Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.768523 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-config-data" (OuterVolumeSpecName: "config-data") pod "11060069-6fb7-4371-9be8-938a92c8269e" (UID: "11060069-6fb7-4371-9be8-938a92c8269e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.769863 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-scripts" (OuterVolumeSpecName: "scripts") pod "11060069-6fb7-4371-9be8-938a92c8269e" (UID: "11060069-6fb7-4371-9be8-938a92c8269e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.793246 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11060069-6fb7-4371-9be8-938a92c8269e-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "11060069-6fb7-4371-9be8-938a92c8269e" (UID: "11060069-6fb7-4371-9be8-938a92c8269e"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.818598 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11060069-6fb7-4371-9be8-938a92c8269e" (UID: "11060069-6fb7-4371-9be8-938a92c8269e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.862332 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.862360 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.862373 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11060069-6fb7-4371-9be8-938a92c8269e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:47:13 crc kubenswrapper[4741]: I0929 20:47:13.862382 4741 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11060069-6fb7-4371-9be8-938a92c8269e-config-data-merged\") on node \"crc\" DevicePath \"\"" Sep 29 20:47:14 crc kubenswrapper[4741]: I0929 20:47:14.240691 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rvkpf" event={"ID":"11060069-6fb7-4371-9be8-938a92c8269e","Type":"ContainerDied","Data":"2e5f410573ef4c26fdd77e2bd8f6b7299aee34fb7dd3acdff66dabb4cd3cfa12"} Sep 29 20:47:14 crc kubenswrapper[4741]: I0929 20:47:14.240739 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e5f410573ef4c26fdd77e2bd8f6b7299aee34fb7dd3acdff66dabb4cd3cfa12" Sep 29 20:47:14 crc kubenswrapper[4741]: I0929 20:47:14.240819 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rvkpf" Sep 29 20:47:16 crc kubenswrapper[4741]: I0929 20:47:16.626802 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-nsg8t" Sep 29 20:47:19 crc kubenswrapper[4741]: I0929 20:47:19.985070 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-v24jm" Sep 29 20:47:21 crc kubenswrapper[4741]: I0929 20:47:21.042353 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-gdqxl"] Sep 29 20:47:21 crc kubenswrapper[4741]: I0929 20:47:21.051700 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-gdqxl"] Sep 29 20:47:21 crc kubenswrapper[4741]: I0929 20:47:21.114570 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f440e0b7-b0d0-4db6-bc89-5173c5ec283f" path="/var/lib/kubelet/pods/f440e0b7-b0d0-4db6-bc89-5173c5ec283f/volumes" Sep 29 20:47:21 crc kubenswrapper[4741]: I0929 20:47:21.687505 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-744s7" Sep 29 20:47:27 crc kubenswrapper[4741]: I0929 20:47:27.976947 4741 scope.go:117] "RemoveContainer" containerID="ebfdd927e0ded53243c1d4d6081b1f11b13a1601727843f2bea4af94847a9764" Sep 29 20:47:31 crc kubenswrapper[4741]: I0929 20:47:31.034158 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d416-account-create-m2jtt"] Sep 29 20:47:31 crc kubenswrapper[4741]: I0929 20:47:31.046765 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d416-account-create-m2jtt"] Sep 29 20:47:31 crc kubenswrapper[4741]: I0929 20:47:31.100012 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aaa33e1-0756-44ef-a69e-cda075bdbeb3" path="/var/lib/kubelet/pods/3aaa33e1-0756-44ef-a69e-cda075bdbeb3/volumes" Sep 29 20:47:31 crc kubenswrapper[4741]: I0929 20:47:31.739183 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:47:31 crc kubenswrapper[4741]: I0929 20:47:31.739236 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:47:37 crc kubenswrapper[4741]: I0929 20:47:37.035770 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-j2k9n"] Sep 29 20:47:37 crc kubenswrapper[4741]: I0929 20:47:37.044995 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-j2k9n"] Sep 29 20:47:37 crc kubenswrapper[4741]: I0929 20:47:37.096836 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04834d12-fe1c-4900-b570-6097aa294692" path="/var/lib/kubelet/pods/04834d12-fe1c-4900-b570-6097aa294692/volumes" Sep 29 20:47:41 crc kubenswrapper[4741]: I0929 20:47:41.683524 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fwn97"] Sep 29 20:47:41 crc kubenswrapper[4741]: I0929 20:47:41.684374 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-fwn97" podUID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerName="octavia-amphora-httpd" containerID="cri-o://56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770" gracePeriod=30 Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.248738 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.356490 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/3c8516c2-d162-47bc-b9f4-bca1708e0868-amphora-image\") pod \"3c8516c2-d162-47bc-b9f4-bca1708e0868\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.356592 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c8516c2-d162-47bc-b9f4-bca1708e0868-httpd-config\") pod \"3c8516c2-d162-47bc-b9f4-bca1708e0868\" (UID: \"3c8516c2-d162-47bc-b9f4-bca1708e0868\") " Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.383247 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c8516c2-d162-47bc-b9f4-bca1708e0868-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "3c8516c2-d162-47bc-b9f4-bca1708e0868" (UID: "3c8516c2-d162-47bc-b9f4-bca1708e0868"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.440903 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c8516c2-d162-47bc-b9f4-bca1708e0868-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "3c8516c2-d162-47bc-b9f4-bca1708e0868" (UID: "3c8516c2-d162-47bc-b9f4-bca1708e0868"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.459844 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c8516c2-d162-47bc-b9f4-bca1708e0868-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.459873 4741 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/3c8516c2-d162-47bc-b9f4-bca1708e0868-amphora-image\") on node \"crc\" DevicePath \"\"" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.517652 4741 generic.go:334] "Generic (PLEG): container finished" podID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerID="56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770" exitCode=0 Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.517692 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fwn97" event={"ID":"3c8516c2-d162-47bc-b9f4-bca1708e0868","Type":"ContainerDied","Data":"56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770"} Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.517717 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fwn97" event={"ID":"3c8516c2-d162-47bc-b9f4-bca1708e0868","Type":"ContainerDied","Data":"19962560e7cae1cd51422ef7430faa6d66870637e1397ef1c37ba27f5c26a119"} Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.517722 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fwn97" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.517734 4741 scope.go:117] "RemoveContainer" containerID="56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.554905 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fwn97"] Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.556253 4741 scope.go:117] "RemoveContainer" containerID="c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.564160 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fwn97"] Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.579016 4741 scope.go:117] "RemoveContainer" containerID="56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770" Sep 29 20:47:42 crc kubenswrapper[4741]: E0929 20:47:42.579368 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770\": container with ID starting with 56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770 not found: ID does not exist" containerID="56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.579536 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770"} err="failed to get container status \"56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770\": rpc error: code = NotFound desc = could not find container \"56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770\": container with ID starting with 56ba18b2db3b6f1ba38dd97bfe11ab309032ff9b6a678fb68f7eaa470b030770 not found: ID does not exist" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.579560 4741 scope.go:117] "RemoveContainer" containerID="c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5" Sep 29 20:47:42 crc kubenswrapper[4741]: E0929 20:47:42.580105 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5\": container with ID starting with c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5 not found: ID does not exist" containerID="c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5" Sep 29 20:47:42 crc kubenswrapper[4741]: I0929 20:47:42.580179 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5"} err="failed to get container status \"c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5\": rpc error: code = NotFound desc = could not find container \"c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5\": container with ID starting with c505d6fe0bd66ba92a7c53dc2dca96868856d2b852b0fbbe9a7ab58af5bbdee5 not found: ID does not exist" Sep 29 20:47:43 crc kubenswrapper[4741]: I0929 20:47:43.120276 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c8516c2-d162-47bc-b9f4-bca1708e0868" path="/var/lib/kubelet/pods/3c8516c2-d162-47bc-b9f4-bca1708e0868/volumes" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.273380 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-vmzlp"] Sep 29 20:47:46 crc kubenswrapper[4741]: E0929 20:47:46.274290 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11060069-6fb7-4371-9be8-938a92c8269e" containerName="octavia-db-sync" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.274303 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="11060069-6fb7-4371-9be8-938a92c8269e" containerName="octavia-db-sync" Sep 29 20:47:46 crc kubenswrapper[4741]: E0929 20:47:46.274319 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerName="octavia-amphora-httpd" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.274325 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerName="octavia-amphora-httpd" Sep 29 20:47:46 crc kubenswrapper[4741]: E0929 20:47:46.274344 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerName="init" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.274350 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerName="init" Sep 29 20:47:46 crc kubenswrapper[4741]: E0929 20:47:46.274361 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11060069-6fb7-4371-9be8-938a92c8269e" containerName="init" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.274366 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="11060069-6fb7-4371-9be8-938a92c8269e" containerName="init" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.274550 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8516c2-d162-47bc-b9f4-bca1708e0868" containerName="octavia-amphora-httpd" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.274573 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="11060069-6fb7-4371-9be8-938a92c8269e" containerName="octavia-db-sync" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.275770 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.279351 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.288521 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-vmzlp"] Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.429802 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/412eba2a-c36e-42a4-a15f-661fbf0d902c-amphora-image\") pod \"octavia-image-upload-59f8cff499-vmzlp\" (UID: \"412eba2a-c36e-42a4-a15f-661fbf0d902c\") " pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.430524 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/412eba2a-c36e-42a4-a15f-661fbf0d902c-httpd-config\") pod \"octavia-image-upload-59f8cff499-vmzlp\" (UID: \"412eba2a-c36e-42a4-a15f-661fbf0d902c\") " pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.532905 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/412eba2a-c36e-42a4-a15f-661fbf0d902c-httpd-config\") pod \"octavia-image-upload-59f8cff499-vmzlp\" (UID: \"412eba2a-c36e-42a4-a15f-661fbf0d902c\") " pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.533041 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/412eba2a-c36e-42a4-a15f-661fbf0d902c-amphora-image\") pod \"octavia-image-upload-59f8cff499-vmzlp\" (UID: \"412eba2a-c36e-42a4-a15f-661fbf0d902c\") " pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.533541 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/412eba2a-c36e-42a4-a15f-661fbf0d902c-amphora-image\") pod \"octavia-image-upload-59f8cff499-vmzlp\" (UID: \"412eba2a-c36e-42a4-a15f-661fbf0d902c\") " pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.538577 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/412eba2a-c36e-42a4-a15f-661fbf0d902c-httpd-config\") pod \"octavia-image-upload-59f8cff499-vmzlp\" (UID: \"412eba2a-c36e-42a4-a15f-661fbf0d902c\") " pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:46 crc kubenswrapper[4741]: I0929 20:47:46.600279 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-vmzlp" Sep 29 20:47:47 crc kubenswrapper[4741]: I0929 20:47:47.063007 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-vmzlp"] Sep 29 20:47:47 crc kubenswrapper[4741]: I0929 20:47:47.559548 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-vmzlp" event={"ID":"412eba2a-c36e-42a4-a15f-661fbf0d902c","Type":"ContainerStarted","Data":"5e4f4b27acb0e59f28928a3c95e81051e7798d088d3b8d260f51233f79fabe25"} Sep 29 20:47:48 crc kubenswrapper[4741]: I0929 20:47:48.572250 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-vmzlp" event={"ID":"412eba2a-c36e-42a4-a15f-661fbf0d902c","Type":"ContainerStarted","Data":"00bc25ae91ae68261fda4c321dab3a36b9049ea5c6b14b30cebd7926f7e4375e"} Sep 29 20:47:49 crc kubenswrapper[4741]: I0929 20:47:49.581153 4741 generic.go:334] "Generic (PLEG): container finished" podID="412eba2a-c36e-42a4-a15f-661fbf0d902c" containerID="00bc25ae91ae68261fda4c321dab3a36b9049ea5c6b14b30cebd7926f7e4375e" exitCode=0 Sep 29 20:47:49 crc kubenswrapper[4741]: I0929 20:47:49.581190 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-vmzlp" event={"ID":"412eba2a-c36e-42a4-a15f-661fbf0d902c","Type":"ContainerDied","Data":"00bc25ae91ae68261fda4c321dab3a36b9049ea5c6b14b30cebd7926f7e4375e"} Sep 29 20:47:51 crc kubenswrapper[4741]: I0929 20:47:51.597721 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-vmzlp" event={"ID":"412eba2a-c36e-42a4-a15f-661fbf0d902c","Type":"ContainerStarted","Data":"080a727255f894128ba0832fed05d90fffcdadaf3ed73cffd0bfbd8547c170e6"} Sep 29 20:47:51 crc kubenswrapper[4741]: I0929 20:47:51.616194 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-vmzlp" podStartSLOduration=1.511041681 podStartE2EDuration="5.61617877s" podCreationTimestamp="2025-09-29 20:47:46 +0000 UTC" firstStartedPulling="2025-09-29 20:47:47.07088829 +0000 UTC m=+5908.718677622" lastFinishedPulling="2025-09-29 20:47:51.176025379 +0000 UTC m=+5912.823814711" observedRunningTime="2025-09-29 20:47:51.60937578 +0000 UTC m=+5913.257165112" watchObservedRunningTime="2025-09-29 20:47:51.61617877 +0000 UTC m=+5913.263968092" Sep 29 20:48:01 crc kubenswrapper[4741]: I0929 20:48:01.738611 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:48:01 crc kubenswrapper[4741]: I0929 20:48:01.739195 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:48:07 crc kubenswrapper[4741]: I0929 20:48:07.048874 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-n8mwn"] Sep 29 20:48:07 crc kubenswrapper[4741]: I0929 20:48:07.058499 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-n8mwn"] Sep 29 20:48:07 crc kubenswrapper[4741]: I0929 20:48:07.098931 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f934fe52-4c4f-45e2-8051-a3e70052e6c6" path="/var/lib/kubelet/pods/f934fe52-4c4f-45e2-8051-a3e70052e6c6/volumes" Sep 29 20:48:17 crc kubenswrapper[4741]: I0929 20:48:17.028627 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3c8d-account-create-4j747"] Sep 29 20:48:17 crc kubenswrapper[4741]: I0929 20:48:17.037138 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3c8d-account-create-4j747"] Sep 29 20:48:17 crc kubenswrapper[4741]: I0929 20:48:17.097801 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91cc136a-99d2-4554-80c9-1e701449f830" path="/var/lib/kubelet/pods/91cc136a-99d2-4554-80c9-1e701449f830/volumes" Sep 29 20:48:26 crc kubenswrapper[4741]: I0929 20:48:26.033051 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-wcpgv"] Sep 29 20:48:26 crc kubenswrapper[4741]: I0929 20:48:26.041703 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-wcpgv"] Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.097212 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc2fd00-e59e-476e-ad96-b2480837eff8" path="/var/lib/kubelet/pods/7fc2fd00-e59e-476e-ad96-b2480837eff8/volumes" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.156613 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-75b885c557-bbtjf"] Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.159310 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.170740 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.171034 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-x9gbm" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.172058 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.172617 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.186081 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-75b885c557-bbtjf"] Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.289805 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.290287 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-log" containerID="cri-o://dc501c03ac22dc0c726eaa49e2b647d24785bbcebab5cc11a0abb939a7413a33" gracePeriod=30 Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.290751 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-httpd" containerID="cri-o://1f348a03ee997223baf508b479abe0f13bfb9f2c307568a6984cda082101b4b6" gracePeriod=30 Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.331985 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a110f476-90be-4557-ae05-9ff2b0777eea-horizon-secret-key\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.332050 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-config-data\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.332090 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-scripts\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.332228 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a110f476-90be-4557-ae05-9ff2b0777eea-logs\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.332446 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h25fd\" (UniqueName: \"kubernetes.io/projected/a110f476-90be-4557-ae05-9ff2b0777eea-kube-api-access-h25fd\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.388811 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-69bcb675c-8jv5j"] Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.390375 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.447680 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a110f476-90be-4557-ae05-9ff2b0777eea-horizon-secret-key\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.447746 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-config-data\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.447772 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-scripts\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.447910 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a110f476-90be-4557-ae05-9ff2b0777eea-logs\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.447970 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h25fd\" (UniqueName: \"kubernetes.io/projected/a110f476-90be-4557-ae05-9ff2b0777eea-kube-api-access-h25fd\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.448538 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a110f476-90be-4557-ae05-9ff2b0777eea-logs\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.449074 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-scripts\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.449949 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-config-data\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.454847 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69bcb675c-8jv5j"] Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.459931 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a110f476-90be-4557-ae05-9ff2b0777eea-horizon-secret-key\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.480266 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h25fd\" (UniqueName: \"kubernetes.io/projected/a110f476-90be-4557-ae05-9ff2b0777eea-kube-api-access-h25fd\") pod \"horizon-75b885c557-bbtjf\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.510121 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.512682 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-log" containerID="cri-o://0af233296b6dbd422d326bf56f0fc7ae5b97f3738b5678eece65c80e7a88d1aa" gracePeriod=30 Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.513114 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-httpd" containerID="cri-o://d85377e8f9ab6570cccf4bf2331260524ea0849da2b65cef38d22408a1f59255" gracePeriod=30 Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.514762 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.550319 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg9dh\" (UniqueName: \"kubernetes.io/projected/a6de352c-a64c-4fae-a629-e630f238cca7-kube-api-access-rg9dh\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.550464 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-config-data\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.550505 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6de352c-a64c-4fae-a629-e630f238cca7-horizon-secret-key\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.550630 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6de352c-a64c-4fae-a629-e630f238cca7-logs\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.550679 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-scripts\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.652302 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6de352c-a64c-4fae-a629-e630f238cca7-logs\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.652353 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-scripts\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.652413 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg9dh\" (UniqueName: \"kubernetes.io/projected/a6de352c-a64c-4fae-a629-e630f238cca7-kube-api-access-rg9dh\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.652463 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-config-data\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.652490 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6de352c-a64c-4fae-a629-e630f238cca7-horizon-secret-key\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.653363 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-scripts\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.653659 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6de352c-a64c-4fae-a629-e630f238cca7-logs\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.654706 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-config-data\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.657461 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6de352c-a64c-4fae-a629-e630f238cca7-horizon-secret-key\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.668952 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg9dh\" (UniqueName: \"kubernetes.io/projected/a6de352c-a64c-4fae-a629-e630f238cca7-kube-api-access-rg9dh\") pod \"horizon-69bcb675c-8jv5j\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.942066 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.977102 4741 generic.go:334] "Generic (PLEG): container finished" podID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerID="dc501c03ac22dc0c726eaa49e2b647d24785bbcebab5cc11a0abb939a7413a33" exitCode=143 Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.977182 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cbf9f732-7174-4fd9-be92-acd5d3571682","Type":"ContainerDied","Data":"dc501c03ac22dc0c726eaa49e2b647d24785bbcebab5cc11a0abb939a7413a33"} Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.983319 4741 generic.go:334] "Generic (PLEG): container finished" podID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerID="0af233296b6dbd422d326bf56f0fc7ae5b97f3738b5678eece65c80e7a88d1aa" exitCode=143 Sep 29 20:48:27 crc kubenswrapper[4741]: I0929 20:48:27.983354 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d875a0-8d33-4254-ad10-0d4d40edfcf9","Type":"ContainerDied","Data":"0af233296b6dbd422d326bf56f0fc7ae5b97f3738b5678eece65c80e7a88d1aa"} Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.018014 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-75b885c557-bbtjf"] Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.086250 4741 scope.go:117] "RemoveContainer" containerID="c9225fd4f7e98d942096d081346767e4a082cb7c8e7dbdad44147cd1c7af0b90" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.092935 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69bcb675c-8jv5j"] Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.130877 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6468bdd6cc-7gxj7"] Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.133043 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.144902 4741 scope.go:117] "RemoveContainer" containerID="0db150efe72d4fbdd7c4b3c21b487efcdc61d047c0ffd28b283643d2508f52a9" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.165584 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6468bdd6cc-7gxj7"] Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.171184 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-config-data\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.171281 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-scripts\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.171315 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-logs\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.171493 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-horizon-secret-key\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.171587 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5fc5\" (UniqueName: \"kubernetes.io/projected/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-kube-api-access-p5fc5\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.196650 4741 scope.go:117] "RemoveContainer" containerID="e6fda1baf2b442d1b416273464fb7067b8431bb7098fb8f6024891751dadfa51" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.248829 4741 scope.go:117] "RemoveContainer" containerID="f721e456d45383581c7a76ef22d3ad312844c3d1ccfe91422acc01d7b1ac0b5b" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.271232 4741 scope.go:117] "RemoveContainer" containerID="e29d5d1ad083a20d0549b54e615438c618b3fc3ffdfe38f0bed341ea57ad555b" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.272830 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-scripts\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.272881 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-logs\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.272956 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-horizon-secret-key\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.273008 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5fc5\" (UniqueName: \"kubernetes.io/projected/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-kube-api-access-p5fc5\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.273054 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-config-data\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.273455 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-logs\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.273559 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-scripts\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.274096 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-config-data\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.284525 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-horizon-secret-key\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.291180 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5fc5\" (UniqueName: \"kubernetes.io/projected/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-kube-api-access-p5fc5\") pod \"horizon-6468bdd6cc-7gxj7\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.462983 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.478754 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69bcb675c-8jv5j"] Sep 29 20:48:28 crc kubenswrapper[4741]: W0929 20:48:28.487193 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6de352c_a64c_4fae_a629_e630f238cca7.slice/crio-047886d51dd9e21aed4addc478c0a5c1f0f364fccfb90502dd7494fe407f51ae WatchSource:0}: Error finding container 047886d51dd9e21aed4addc478c0a5c1f0f364fccfb90502dd7494fe407f51ae: Status 404 returned error can't find the container with id 047886d51dd9e21aed4addc478c0a5c1f0f364fccfb90502dd7494fe407f51ae Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.906734 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6468bdd6cc-7gxj7"] Sep 29 20:48:28 crc kubenswrapper[4741]: W0929 20:48:28.926695 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aa2eebe_39ed_478d_84ca_f2fd4d1b8f15.slice/crio-147b2a17add310dc1ba9698a01d60668e61545de3ed3681619e4c89de726da3f WatchSource:0}: Error finding container 147b2a17add310dc1ba9698a01d60668e61545de3ed3681619e4c89de726da3f: Status 404 returned error can't find the container with id 147b2a17add310dc1ba9698a01d60668e61545de3ed3681619e4c89de726da3f Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.997039 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69bcb675c-8jv5j" event={"ID":"a6de352c-a64c-4fae-a629-e630f238cca7","Type":"ContainerStarted","Data":"047886d51dd9e21aed4addc478c0a5c1f0f364fccfb90502dd7494fe407f51ae"} Sep 29 20:48:28 crc kubenswrapper[4741]: I0929 20:48:28.999717 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75b885c557-bbtjf" event={"ID":"a110f476-90be-4557-ae05-9ff2b0777eea","Type":"ContainerStarted","Data":"aebaf033a1d0139bb49b06210992dc609d54c399c2aff833b9394ee160f271d9"} Sep 29 20:48:29 crc kubenswrapper[4741]: I0929 20:48:29.001083 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6468bdd6cc-7gxj7" event={"ID":"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15","Type":"ContainerStarted","Data":"147b2a17add310dc1ba9698a01d60668e61545de3ed3681619e4c89de726da3f"} Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.023153 4741 generic.go:334] "Generic (PLEG): container finished" podID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerID="d85377e8f9ab6570cccf4bf2331260524ea0849da2b65cef38d22408a1f59255" exitCode=0 Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.023275 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d875a0-8d33-4254-ad10-0d4d40edfcf9","Type":"ContainerDied","Data":"d85377e8f9ab6570cccf4bf2331260524ea0849da2b65cef38d22408a1f59255"} Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.026246 4741 generic.go:334] "Generic (PLEG): container finished" podID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerID="1f348a03ee997223baf508b479abe0f13bfb9f2c307568a6984cda082101b4b6" exitCode=0 Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.026291 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cbf9f732-7174-4fd9-be92-acd5d3571682","Type":"ContainerDied","Data":"1f348a03ee997223baf508b479abe0f13bfb9f2c307568a6984cda082101b4b6"} Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.057206 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.1.38:9292/healthcheck\": dial tcp 10.217.1.38:9292: connect: connection refused" Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.057220 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.1.38:9292/healthcheck\": dial tcp 10.217.1.38:9292: connect: connection refused" Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.738676 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.738729 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.738769 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.739223 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:48:31 crc kubenswrapper[4741]: I0929 20:48:31.739266 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" gracePeriod=600 Sep 29 20:48:32 crc kubenswrapper[4741]: I0929 20:48:32.038435 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" exitCode=0 Sep 29 20:48:32 crc kubenswrapper[4741]: I0929 20:48:32.038514 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7"} Sep 29 20:48:32 crc kubenswrapper[4741]: I0929 20:48:32.038759 4741 scope.go:117] "RemoveContainer" containerID="564907e3fd83739b2ba2e94f48864f64d35270919df4a735eaebeff724bee01a" Sep 29 20:48:34 crc kubenswrapper[4741]: I0929 20:48:34.074992 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.1.39:9292/healthcheck\": dial tcp 10.217.1.39:9292: connect: connection refused" Sep 29 20:48:34 crc kubenswrapper[4741]: I0929 20:48:34.075179 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.1.39:9292/healthcheck\": dial tcp 10.217.1.39:9292: connect: connection refused" Sep 29 20:48:35 crc kubenswrapper[4741]: E0929 20:48:35.018570 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.083651 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:48:35 crc kubenswrapper[4741]: E0929 20:48:35.086300 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.493784 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.528608 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-combined-ca-bundle\") pod \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.528759 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-ceph\") pod \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.528896 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64h9v\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-kube-api-access-64h9v\") pod \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.528973 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-scripts\") pod \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.529004 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-httpd-run\") pod \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.529096 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-logs\") pod \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.529129 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-config-data\") pod \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\" (UID: \"e2d875a0-8d33-4254-ad10-0d4d40edfcf9\") " Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.530537 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e2d875a0-8d33-4254-ad10-0d4d40edfcf9" (UID: "e2d875a0-8d33-4254-ad10-0d4d40edfcf9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.530819 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-logs" (OuterVolumeSpecName: "logs") pod "e2d875a0-8d33-4254-ad10-0d4d40edfcf9" (UID: "e2d875a0-8d33-4254-ad10-0d4d40edfcf9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.533784 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-scripts" (OuterVolumeSpecName: "scripts") pod "e2d875a0-8d33-4254-ad10-0d4d40edfcf9" (UID: "e2d875a0-8d33-4254-ad10-0d4d40edfcf9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.543364 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-kube-api-access-64h9v" (OuterVolumeSpecName: "kube-api-access-64h9v") pod "e2d875a0-8d33-4254-ad10-0d4d40edfcf9" (UID: "e2d875a0-8d33-4254-ad10-0d4d40edfcf9"). InnerVolumeSpecName "kube-api-access-64h9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.546280 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-ceph" (OuterVolumeSpecName: "ceph") pod "e2d875a0-8d33-4254-ad10-0d4d40edfcf9" (UID: "e2d875a0-8d33-4254-ad10-0d4d40edfcf9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.588859 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2d875a0-8d33-4254-ad10-0d4d40edfcf9" (UID: "e2d875a0-8d33-4254-ad10-0d4d40edfcf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.631832 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.631861 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.631872 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.631880 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64h9v\" (UniqueName: \"kubernetes.io/projected/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-kube-api-access-64h9v\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.631889 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.631897 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.653084 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-config-data" (OuterVolumeSpecName: "config-data") pod "e2d875a0-8d33-4254-ad10-0d4d40edfcf9" (UID: "e2d875a0-8d33-4254-ad10-0d4d40edfcf9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:48:35 crc kubenswrapper[4741]: I0929 20:48:35.733840 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d875a0-8d33-4254-ad10-0d4d40edfcf9-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.094993 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69bcb675c-8jv5j" event={"ID":"a6de352c-a64c-4fae-a629-e630f238cca7","Type":"ContainerStarted","Data":"3f85de8aa06c465a1b83de0f3fbc707cb8fbd18be26535268631c730ee3df03b"} Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.095198 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69bcb675c-8jv5j" event={"ID":"a6de352c-a64c-4fae-a629-e630f238cca7","Type":"ContainerStarted","Data":"ab7f96a2edfd8c6e4d132fe5747de69e0f613625be7216e7747dc2445e1655c6"} Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.095248 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69bcb675c-8jv5j" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon" containerID="cri-o://3f85de8aa06c465a1b83de0f3fbc707cb8fbd18be26535268631c730ee3df03b" gracePeriod=30 Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.095557 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69bcb675c-8jv5j" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon-log" containerID="cri-o://ab7f96a2edfd8c6e4d132fe5747de69e0f613625be7216e7747dc2445e1655c6" gracePeriod=30 Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.100015 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d875a0-8d33-4254-ad10-0d4d40edfcf9","Type":"ContainerDied","Data":"8d0a22297d00c3f5e6d7ca966f2c22c1a18eabca53d66a5e0308631222f3c224"} Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.100065 4741 scope.go:117] "RemoveContainer" containerID="d85377e8f9ab6570cccf4bf2331260524ea0849da2b65cef38d22408a1f59255" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.100152 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.103161 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6468bdd6cc-7gxj7" event={"ID":"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15","Type":"ContainerStarted","Data":"30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4"} Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.103188 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6468bdd6cc-7gxj7" event={"ID":"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15","Type":"ContainerStarted","Data":"d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58"} Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.107278 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75b885c557-bbtjf" event={"ID":"a110f476-90be-4557-ae05-9ff2b0777eea","Type":"ContainerStarted","Data":"f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e"} Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.107309 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75b885c557-bbtjf" event={"ID":"a110f476-90be-4557-ae05-9ff2b0777eea","Type":"ContainerStarted","Data":"f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb"} Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.122673 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-69bcb675c-8jv5j" podStartSLOduration=2.382215104 podStartE2EDuration="9.122650811s" podCreationTimestamp="2025-09-29 20:48:27 +0000 UTC" firstStartedPulling="2025-09-29 20:48:28.494789615 +0000 UTC m=+5950.142578947" lastFinishedPulling="2025-09-29 20:48:35.235225322 +0000 UTC m=+5956.883014654" observedRunningTime="2025-09-29 20:48:36.115551443 +0000 UTC m=+5957.763340775" watchObservedRunningTime="2025-09-29 20:48:36.122650811 +0000 UTC m=+5957.770440143" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.146866 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-75b885c557-bbtjf" podStartSLOduration=1.991356763 podStartE2EDuration="9.146846542s" podCreationTimestamp="2025-09-29 20:48:27 +0000 UTC" firstStartedPulling="2025-09-29 20:48:28.039335712 +0000 UTC m=+5949.687125054" lastFinishedPulling="2025-09-29 20:48:35.194825501 +0000 UTC m=+5956.842614833" observedRunningTime="2025-09-29 20:48:36.140675942 +0000 UTC m=+5957.788465264" watchObservedRunningTime="2025-09-29 20:48:36.146846542 +0000 UTC m=+5957.794635874" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.147911 4741 scope.go:117] "RemoveContainer" containerID="0af233296b6dbd422d326bf56f0fc7ae5b97f3738b5678eece65c80e7a88d1aa" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.170670 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6468bdd6cc-7gxj7" podStartSLOduration=1.904322013 podStartE2EDuration="8.170653148s" podCreationTimestamp="2025-09-29 20:48:28 +0000 UTC" firstStartedPulling="2025-09-29 20:48:28.928643631 +0000 UTC m=+5950.576432963" lastFinishedPulling="2025-09-29 20:48:35.194974766 +0000 UTC m=+5956.842764098" observedRunningTime="2025-09-29 20:48:36.168521239 +0000 UTC m=+5957.816310571" watchObservedRunningTime="2025-09-29 20:48:36.170653148 +0000 UTC m=+5957.818442480" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.215117 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.229153 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.242554 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:48:36 crc kubenswrapper[4741]: E0929 20:48:36.242970 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-httpd" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.242985 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-httpd" Sep 29 20:48:36 crc kubenswrapper[4741]: E0929 20:48:36.243008 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-log" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.243014 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-log" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.243264 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-log" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.243286 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" containerName="glance-httpd" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.244219 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.247104 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.251605 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.352153 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f9554882-6011-402e-b598-e15f0284c296-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.352227 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmrhq\" (UniqueName: \"kubernetes.io/projected/f9554882-6011-402e-b598-e15f0284c296-kube-api-access-zmrhq\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.352593 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9554882-6011-402e-b598-e15f0284c296-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.352651 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.352713 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.352746 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9554882-6011-402e-b598-e15f0284c296-logs\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.352779 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.411903 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.457994 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-ceph\") pod \"cbf9f732-7174-4fd9-be92-acd5d3571682\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.458117 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-combined-ca-bundle\") pod \"cbf9f732-7174-4fd9-be92-acd5d3571682\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.458165 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-scripts\") pod \"cbf9f732-7174-4fd9-be92-acd5d3571682\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.458251 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-httpd-run\") pod \"cbf9f732-7174-4fd9-be92-acd5d3571682\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.458292 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnrr4\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-kube-api-access-gnrr4\") pod \"cbf9f732-7174-4fd9-be92-acd5d3571682\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.458337 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-logs\") pod \"cbf9f732-7174-4fd9-be92-acd5d3571682\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.458427 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-config-data\") pod \"cbf9f732-7174-4fd9-be92-acd5d3571682\" (UID: \"cbf9f732-7174-4fd9-be92-acd5d3571682\") " Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.459047 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f9554882-6011-402e-b598-e15f0284c296-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.459083 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmrhq\" (UniqueName: \"kubernetes.io/projected/f9554882-6011-402e-b598-e15f0284c296-kube-api-access-zmrhq\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.459214 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9554882-6011-402e-b598-e15f0284c296-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.459252 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.459281 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.459303 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9554882-6011-402e-b598-e15f0284c296-logs\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.459325 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.464254 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9554882-6011-402e-b598-e15f0284c296-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.470993 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9554882-6011-402e-b598-e15f0284c296-logs\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.471657 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-logs" (OuterVolumeSpecName: "logs") pod "cbf9f732-7174-4fd9-be92-acd5d3571682" (UID: "cbf9f732-7174-4fd9-be92-acd5d3571682"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.472148 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cbf9f732-7174-4fd9-be92-acd5d3571682" (UID: "cbf9f732-7174-4fd9-be92-acd5d3571682"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.477597 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f9554882-6011-402e-b598-e15f0284c296-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.492620 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.502627 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-ceph" (OuterVolumeSpecName: "ceph") pod "cbf9f732-7174-4fd9-be92-acd5d3571682" (UID: "cbf9f732-7174-4fd9-be92-acd5d3571682"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.503256 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-scripts" (OuterVolumeSpecName: "scripts") pod "cbf9f732-7174-4fd9-be92-acd5d3571682" (UID: "cbf9f732-7174-4fd9-be92-acd5d3571682"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.503354 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmrhq\" (UniqueName: \"kubernetes.io/projected/f9554882-6011-402e-b598-e15f0284c296-kube-api-access-zmrhq\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.505843 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-kube-api-access-gnrr4" (OuterVolumeSpecName: "kube-api-access-gnrr4") pod "cbf9f732-7174-4fd9-be92-acd5d3571682" (UID: "cbf9f732-7174-4fd9-be92-acd5d3571682"). InnerVolumeSpecName "kube-api-access-gnrr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.520383 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.528062 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9554882-6011-402e-b598-e15f0284c296-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f9554882-6011-402e-b598-e15f0284c296\") " pod="openstack/glance-default-internal-api-0" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.537663 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbf9f732-7174-4fd9-be92-acd5d3571682" (UID: "cbf9f732-7174-4fd9-be92-acd5d3571682"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.542528 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-config-data" (OuterVolumeSpecName: "config-data") pod "cbf9f732-7174-4fd9-be92-acd5d3571682" (UID: "cbf9f732-7174-4fd9-be92-acd5d3571682"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.562813 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.562841 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.562853 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.562861 4741 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.562870 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnrr4\" (UniqueName: \"kubernetes.io/projected/cbf9f732-7174-4fd9-be92-acd5d3571682-kube-api-access-gnrr4\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.562878 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbf9f732-7174-4fd9-be92-acd5d3571682-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.562888 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf9f732-7174-4fd9-be92-acd5d3571682-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:48:36 crc kubenswrapper[4741]: I0929 20:48:36.569383 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.099271 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d875a0-8d33-4254-ad10-0d4d40edfcf9" path="/var/lib/kubelet/pods/e2d875a0-8d33-4254-ad10-0d4d40edfcf9/volumes" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.117656 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cbf9f732-7174-4fd9-be92-acd5d3571682","Type":"ContainerDied","Data":"58a38e3d293df93a9d5f2ee7002c03aeaad4eac75da5fc1a84ee09c15a40f433"} Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.117714 4741 scope.go:117] "RemoveContainer" containerID="1f348a03ee997223baf508b479abe0f13bfb9f2c307568a6984cda082101b4b6" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.117923 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.157925 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.177596 4741 scope.go:117] "RemoveContainer" containerID="dc501c03ac22dc0c726eaa49e2b647d24785bbcebab5cc11a0abb939a7413a33" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.182855 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:48:37 crc kubenswrapper[4741]: W0929 20:48:37.184886 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9554882_6011_402e_b598_e15f0284c296.slice/crio-d8f9bc81fb89afe1ed006e707e25b5dbdb93d3064e486c6a581695a0e738b737 WatchSource:0}: Error finding container d8f9bc81fb89afe1ed006e707e25b5dbdb93d3064e486c6a581695a0e738b737: Status 404 returned error can't find the container with id d8f9bc81fb89afe1ed006e707e25b5dbdb93d3064e486c6a581695a0e738b737 Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.209294 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.220891 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:48:37 crc kubenswrapper[4741]: E0929 20:48:37.221398 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-log" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.221440 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-log" Sep 29 20:48:37 crc kubenswrapper[4741]: E0929 20:48:37.221454 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-httpd" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.221461 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-httpd" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.221635 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-log" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.221656 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" containerName="glance-httpd" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.222774 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.224795 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.228559 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.282973 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnzrq\" (UniqueName: \"kubernetes.io/projected/c5bf2f12-6a59-4aee-84f5-62c964edab86-kube-api-access-hnzrq\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.283018 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5bf2f12-6a59-4aee-84f5-62c964edab86-ceph\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.283044 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5bf2f12-6a59-4aee-84f5-62c964edab86-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.283069 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.283159 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.283197 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bf2f12-6a59-4aee-84f5-62c964edab86-logs\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.283254 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.385804 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.385865 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bf2f12-6a59-4aee-84f5-62c964edab86-logs\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.385921 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.385963 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnzrq\" (UniqueName: \"kubernetes.io/projected/c5bf2f12-6a59-4aee-84f5-62c964edab86-kube-api-access-hnzrq\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.385986 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5bf2f12-6a59-4aee-84f5-62c964edab86-ceph\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.386005 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5bf2f12-6a59-4aee-84f5-62c964edab86-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.386028 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.387068 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5bf2f12-6a59-4aee-84f5-62c964edab86-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.387435 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bf2f12-6a59-4aee-84f5-62c964edab86-logs\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.402022 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5bf2f12-6a59-4aee-84f5-62c964edab86-ceph\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.402240 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.402262 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.402860 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bf2f12-6a59-4aee-84f5-62c964edab86-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.405972 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnzrq\" (UniqueName: \"kubernetes.io/projected/c5bf2f12-6a59-4aee-84f5-62c964edab86-kube-api-access-hnzrq\") pod \"glance-default-external-api-0\" (UID: \"c5bf2f12-6a59-4aee-84f5-62c964edab86\") " pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.521464 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.521561 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.554441 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 20:48:37 crc kubenswrapper[4741]: I0929 20:48:37.943610 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:48:38 crc kubenswrapper[4741]: I0929 20:48:38.109334 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 20:48:38 crc kubenswrapper[4741]: W0929 20:48:38.118039 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5bf2f12_6a59_4aee_84f5_62c964edab86.slice/crio-0712506b40258a5b92a2ac1f6e0316cfdec5b2b6776c226251300a10ef76978c WatchSource:0}: Error finding container 0712506b40258a5b92a2ac1f6e0316cfdec5b2b6776c226251300a10ef76978c: Status 404 returned error can't find the container with id 0712506b40258a5b92a2ac1f6e0316cfdec5b2b6776c226251300a10ef76978c Sep 29 20:48:38 crc kubenswrapper[4741]: I0929 20:48:38.138790 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f9554882-6011-402e-b598-e15f0284c296","Type":"ContainerStarted","Data":"1efb0b92031bb89cd23be5c5b295fe8ab8eb6505e90eeacfe9f9eff09f40a431"} Sep 29 20:48:38 crc kubenswrapper[4741]: I0929 20:48:38.138828 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f9554882-6011-402e-b598-e15f0284c296","Type":"ContainerStarted","Data":"d8f9bc81fb89afe1ed006e707e25b5dbdb93d3064e486c6a581695a0e738b737"} Sep 29 20:48:38 crc kubenswrapper[4741]: I0929 20:48:38.140502 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5bf2f12-6a59-4aee-84f5-62c964edab86","Type":"ContainerStarted","Data":"0712506b40258a5b92a2ac1f6e0316cfdec5b2b6776c226251300a10ef76978c"} Sep 29 20:48:38 crc kubenswrapper[4741]: I0929 20:48:38.464068 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:38 crc kubenswrapper[4741]: I0929 20:48:38.464512 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:48:39 crc kubenswrapper[4741]: I0929 20:48:39.112464 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbf9f732-7174-4fd9-be92-acd5d3571682" path="/var/lib/kubelet/pods/cbf9f732-7174-4fd9-be92-acd5d3571682/volumes" Sep 29 20:48:39 crc kubenswrapper[4741]: I0929 20:48:39.152005 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f9554882-6011-402e-b598-e15f0284c296","Type":"ContainerStarted","Data":"41c4ec6a8a85a80ffe935cdc516f76497b39a43196e9f1ee464388d3de6b07c2"} Sep 29 20:48:39 crc kubenswrapper[4741]: I0929 20:48:39.154036 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5bf2f12-6a59-4aee-84f5-62c964edab86","Type":"ContainerStarted","Data":"2a3b6dd2443b3fb7901d7f82cbe81c22976f1b66865da72762f43347d1c1c35f"} Sep 29 20:48:39 crc kubenswrapper[4741]: I0929 20:48:39.172898 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.172880917 podStartE2EDuration="3.172880917s" podCreationTimestamp="2025-09-29 20:48:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:48:39.172697161 +0000 UTC m=+5960.820486493" watchObservedRunningTime="2025-09-29 20:48:39.172880917 +0000 UTC m=+5960.820670249" Sep 29 20:48:40 crc kubenswrapper[4741]: I0929 20:48:40.167678 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5bf2f12-6a59-4aee-84f5-62c964edab86","Type":"ContainerStarted","Data":"bfec4b09cbdbd341a0e2811de6f8913561b1a5f74600c5cee7cee17b12c45524"} Sep 29 20:48:40 crc kubenswrapper[4741]: I0929 20:48:40.194310 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.194291783 podStartE2EDuration="3.194291783s" podCreationTimestamp="2025-09-29 20:48:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:48:40.191910386 +0000 UTC m=+5961.839699738" watchObservedRunningTime="2025-09-29 20:48:40.194291783 +0000 UTC m=+5961.842081115" Sep 29 20:48:46 crc kubenswrapper[4741]: I0929 20:48:46.570706 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:46 crc kubenswrapper[4741]: I0929 20:48:46.571274 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:46 crc kubenswrapper[4741]: I0929 20:48:46.607113 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:46 crc kubenswrapper[4741]: I0929 20:48:46.621169 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:47 crc kubenswrapper[4741]: I0929 20:48:47.240072 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:47 crc kubenswrapper[4741]: I0929 20:48:47.240381 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:47 crc kubenswrapper[4741]: I0929 20:48:47.516952 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-75b885c557-bbtjf" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Sep 29 20:48:47 crc kubenswrapper[4741]: I0929 20:48:47.555324 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 20:48:47 crc kubenswrapper[4741]: I0929 20:48:47.556359 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 20:48:47 crc kubenswrapper[4741]: I0929 20:48:47.597094 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 20:48:47 crc kubenswrapper[4741]: I0929 20:48:47.601128 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 20:48:48 crc kubenswrapper[4741]: I0929 20:48:48.085693 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:48:48 crc kubenswrapper[4741]: E0929 20:48:48.086146 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:48:48 crc kubenswrapper[4741]: I0929 20:48:48.277708 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 20:48:48 crc kubenswrapper[4741]: I0929 20:48:48.277773 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 20:48:48 crc kubenswrapper[4741]: I0929 20:48:48.465813 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6468bdd6cc-7gxj7" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Sep 29 20:48:49 crc kubenswrapper[4741]: I0929 20:48:49.233729 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:49 crc kubenswrapper[4741]: I0929 20:48:49.238503 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 20:48:50 crc kubenswrapper[4741]: I0929 20:48:50.231747 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 20:48:50 crc kubenswrapper[4741]: I0929 20:48:50.291327 4741 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 20:48:50 crc kubenswrapper[4741]: I0929 20:48:50.404610 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 20:48:59 crc kubenswrapper[4741]: I0929 20:48:59.374128 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:49:00 crc kubenswrapper[4741]: I0929 20:49:00.085425 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:49:00 crc kubenswrapper[4741]: E0929 20:49:00.086157 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:49:00 crc kubenswrapper[4741]: I0929 20:49:00.231112 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:49:01 crc kubenswrapper[4741]: I0929 20:49:01.022622 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:49:01 crc kubenswrapper[4741]: I0929 20:49:01.908233 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:49:01 crc kubenswrapper[4741]: I0929 20:49:01.990999 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75b885c557-bbtjf"] Sep 29 20:49:01 crc kubenswrapper[4741]: I0929 20:49:01.993236 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75b885c557-bbtjf" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon-log" containerID="cri-o://f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb" gracePeriod=30 Sep 29 20:49:01 crc kubenswrapper[4741]: I0929 20:49:01.993367 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75b885c557-bbtjf" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" containerID="cri-o://f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e" gracePeriod=30 Sep 29 20:49:05 crc kubenswrapper[4741]: I0929 20:49:05.454549 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75b885c557-bbtjf" event={"ID":"a110f476-90be-4557-ae05-9ff2b0777eea","Type":"ContainerDied","Data":"f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e"} Sep 29 20:49:05 crc kubenswrapper[4741]: I0929 20:49:05.454512 4741 generic.go:334] "Generic (PLEG): container finished" podID="a110f476-90be-4557-ae05-9ff2b0777eea" containerID="f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e" exitCode=0 Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.469905 4741 generic.go:334] "Generic (PLEG): container finished" podID="a6de352c-a64c-4fae-a629-e630f238cca7" containerID="3f85de8aa06c465a1b83de0f3fbc707cb8fbd18be26535268631c730ee3df03b" exitCode=137 Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.470184 4741 generic.go:334] "Generic (PLEG): container finished" podID="a6de352c-a64c-4fae-a629-e630f238cca7" containerID="ab7f96a2edfd8c6e4d132fe5747de69e0f613625be7216e7747dc2445e1655c6" exitCode=137 Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.469979 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69bcb675c-8jv5j" event={"ID":"a6de352c-a64c-4fae-a629-e630f238cca7","Type":"ContainerDied","Data":"3f85de8aa06c465a1b83de0f3fbc707cb8fbd18be26535268631c730ee3df03b"} Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.470214 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69bcb675c-8jv5j" event={"ID":"a6de352c-a64c-4fae-a629-e630f238cca7","Type":"ContainerDied","Data":"ab7f96a2edfd8c6e4d132fe5747de69e0f613625be7216e7747dc2445e1655c6"} Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.562892 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.725925 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-scripts\") pod \"a6de352c-a64c-4fae-a629-e630f238cca7\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.726297 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg9dh\" (UniqueName: \"kubernetes.io/projected/a6de352c-a64c-4fae-a629-e630f238cca7-kube-api-access-rg9dh\") pod \"a6de352c-a64c-4fae-a629-e630f238cca7\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.726773 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6de352c-a64c-4fae-a629-e630f238cca7-logs" (OuterVolumeSpecName: "logs") pod "a6de352c-a64c-4fae-a629-e630f238cca7" (UID: "a6de352c-a64c-4fae-a629-e630f238cca7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.726854 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6de352c-a64c-4fae-a629-e630f238cca7-logs\") pod \"a6de352c-a64c-4fae-a629-e630f238cca7\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.726997 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6de352c-a64c-4fae-a629-e630f238cca7-horizon-secret-key\") pod \"a6de352c-a64c-4fae-a629-e630f238cca7\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.727051 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-config-data\") pod \"a6de352c-a64c-4fae-a629-e630f238cca7\" (UID: \"a6de352c-a64c-4fae-a629-e630f238cca7\") " Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.727508 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6de352c-a64c-4fae-a629-e630f238cca7-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.735583 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6de352c-a64c-4fae-a629-e630f238cca7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a6de352c-a64c-4fae-a629-e630f238cca7" (UID: "a6de352c-a64c-4fae-a629-e630f238cca7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.736262 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6de352c-a64c-4fae-a629-e630f238cca7-kube-api-access-rg9dh" (OuterVolumeSpecName: "kube-api-access-rg9dh") pod "a6de352c-a64c-4fae-a629-e630f238cca7" (UID: "a6de352c-a64c-4fae-a629-e630f238cca7"). InnerVolumeSpecName "kube-api-access-rg9dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.756148 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-scripts" (OuterVolumeSpecName: "scripts") pod "a6de352c-a64c-4fae-a629-e630f238cca7" (UID: "a6de352c-a64c-4fae-a629-e630f238cca7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.760297 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-config-data" (OuterVolumeSpecName: "config-data") pod "a6de352c-a64c-4fae-a629-e630f238cca7" (UID: "a6de352c-a64c-4fae-a629-e630f238cca7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.829229 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.829258 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6de352c-a64c-4fae-a629-e630f238cca7-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.829270 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg9dh\" (UniqueName: \"kubernetes.io/projected/a6de352c-a64c-4fae-a629-e630f238cca7-kube-api-access-rg9dh\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:06 crc kubenswrapper[4741]: I0929 20:49:06.829281 4741 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6de352c-a64c-4fae-a629-e630f238cca7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:07 crc kubenswrapper[4741]: I0929 20:49:07.483550 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69bcb675c-8jv5j" event={"ID":"a6de352c-a64c-4fae-a629-e630f238cca7","Type":"ContainerDied","Data":"047886d51dd9e21aed4addc478c0a5c1f0f364fccfb90502dd7494fe407f51ae"} Sep 29 20:49:07 crc kubenswrapper[4741]: I0929 20:49:07.483605 4741 scope.go:117] "RemoveContainer" containerID="3f85de8aa06c465a1b83de0f3fbc707cb8fbd18be26535268631c730ee3df03b" Sep 29 20:49:07 crc kubenswrapper[4741]: I0929 20:49:07.484651 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69bcb675c-8jv5j" Sep 29 20:49:07 crc kubenswrapper[4741]: I0929 20:49:07.512219 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69bcb675c-8jv5j"] Sep 29 20:49:07 crc kubenswrapper[4741]: I0929 20:49:07.516284 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-75b885c557-bbtjf" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Sep 29 20:49:07 crc kubenswrapper[4741]: I0929 20:49:07.524825 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-69bcb675c-8jv5j"] Sep 29 20:49:07 crc kubenswrapper[4741]: I0929 20:49:07.673179 4741 scope.go:117] "RemoveContainer" containerID="ab7f96a2edfd8c6e4d132fe5747de69e0f613625be7216e7747dc2445e1655c6" Sep 29 20:49:08 crc kubenswrapper[4741]: I0929 20:49:08.035059 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zln86"] Sep 29 20:49:08 crc kubenswrapper[4741]: I0929 20:49:08.042867 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zln86"] Sep 29 20:49:09 crc kubenswrapper[4741]: I0929 20:49:09.098199 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" path="/var/lib/kubelet/pods/a6de352c-a64c-4fae-a629-e630f238cca7/volumes" Sep 29 20:49:09 crc kubenswrapper[4741]: I0929 20:49:09.098953 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfbf2271-9d58-45c6-9e0b-961036c2c627" path="/var/lib/kubelet/pods/dfbf2271-9d58-45c6-9e0b-961036c2c627/volumes" Sep 29 20:49:15 crc kubenswrapper[4741]: I0929 20:49:15.086650 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:49:15 crc kubenswrapper[4741]: E0929 20:49:15.087343 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:49:17 crc kubenswrapper[4741]: I0929 20:49:17.515901 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-75b885c557-bbtjf" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Sep 29 20:49:18 crc kubenswrapper[4741]: I0929 20:49:18.038523 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-4c9d-account-create-f9lfc"] Sep 29 20:49:18 crc kubenswrapper[4741]: I0929 20:49:18.049642 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-4c9d-account-create-f9lfc"] Sep 29 20:49:19 crc kubenswrapper[4741]: I0929 20:49:19.105822 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b6532e-5bff-48f4-8b82-78d6eb89eeba" path="/var/lib/kubelet/pods/37b6532e-5bff-48f4-8b82-78d6eb89eeba/volumes" Sep 29 20:49:26 crc kubenswrapper[4741]: I0929 20:49:26.033020 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-rbmn8"] Sep 29 20:49:26 crc kubenswrapper[4741]: I0929 20:49:26.043035 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-rbmn8"] Sep 29 20:49:27 crc kubenswrapper[4741]: I0929 20:49:27.100560 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6558a7e-b784-4232-999e-657db5dccd23" path="/var/lib/kubelet/pods/f6558a7e-b784-4232-999e-657db5dccd23/volumes" Sep 29 20:49:27 crc kubenswrapper[4741]: I0929 20:49:27.515939 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-75b885c557-bbtjf" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Sep 29 20:49:27 crc kubenswrapper[4741]: I0929 20:49:27.516133 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:49:28 crc kubenswrapper[4741]: I0929 20:49:28.429982 4741 scope.go:117] "RemoveContainer" containerID="262f21044f2140d3cf9060091c75dec8287e72aeeca82e207e092274a3428c8e" Sep 29 20:49:28 crc kubenswrapper[4741]: I0929 20:49:28.480190 4741 scope.go:117] "RemoveContainer" containerID="c763d816c0f59d938e26f8ec0ba6aff3a57c1a8a012e795e7797d144cbc1f8ea" Sep 29 20:49:28 crc kubenswrapper[4741]: I0929 20:49:28.519463 4741 scope.go:117] "RemoveContainer" containerID="ced716cb51e4174f4da4df1c9b3e22239cb5f0db87b81e7c5bb10148cfaa27b8" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.092908 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:49:29 crc kubenswrapper[4741]: E0929 20:49:29.093368 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.895448 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ph5ql"] Sep 29 20:49:29 crc kubenswrapper[4741]: E0929 20:49:29.895917 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon-log" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.895930 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon-log" Sep 29 20:49:29 crc kubenswrapper[4741]: E0929 20:49:29.895942 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.895950 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.896150 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon-log" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.896162 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6de352c-a64c-4fae-a629-e630f238cca7" containerName="horizon" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.897600 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:29 crc kubenswrapper[4741]: I0929 20:49:29.913598 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ph5ql"] Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.003665 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4q76\" (UniqueName: \"kubernetes.io/projected/ff59d7a0-7c46-4595-9078-499ce72b7b7f-kube-api-access-n4q76\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.003708 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-utilities\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.003751 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-catalog-content\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.105238 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4q76\" (UniqueName: \"kubernetes.io/projected/ff59d7a0-7c46-4595-9078-499ce72b7b7f-kube-api-access-n4q76\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.105283 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-utilities\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.105325 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-catalog-content\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.105850 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-utilities\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.105895 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-catalog-content\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.131899 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4q76\" (UniqueName: \"kubernetes.io/projected/ff59d7a0-7c46-4595-9078-499ce72b7b7f-kube-api-access-n4q76\") pod \"redhat-marketplace-ph5ql\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.240861 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.712252 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ph5ql"] Sep 29 20:49:30 crc kubenswrapper[4741]: I0929 20:49:30.776860 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ph5ql" event={"ID":"ff59d7a0-7c46-4595-9078-499ce72b7b7f","Type":"ContainerStarted","Data":"d1ffeb121c6ee79f2877f4c2b551d29c72a033acd9f6a2ac9f1e5f13b9d0b054"} Sep 29 20:49:31 crc kubenswrapper[4741]: I0929 20:49:31.787804 4741 generic.go:334] "Generic (PLEG): container finished" podID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerID="3d3b48e4ccc8051edcbf97afbf91c86def81c96532f0c9fe5e3a1987e130076b" exitCode=0 Sep 29 20:49:31 crc kubenswrapper[4741]: I0929 20:49:31.787887 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ph5ql" event={"ID":"ff59d7a0-7c46-4595-9078-499ce72b7b7f","Type":"ContainerDied","Data":"3d3b48e4ccc8051edcbf97afbf91c86def81c96532f0c9fe5e3a1987e130076b"} Sep 29 20:49:31 crc kubenswrapper[4741]: I0929 20:49:31.791416 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.438074 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.561077 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a110f476-90be-4557-ae05-9ff2b0777eea-logs\") pod \"a110f476-90be-4557-ae05-9ff2b0777eea\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.561145 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-scripts\") pod \"a110f476-90be-4557-ae05-9ff2b0777eea\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.561275 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h25fd\" (UniqueName: \"kubernetes.io/projected/a110f476-90be-4557-ae05-9ff2b0777eea-kube-api-access-h25fd\") pod \"a110f476-90be-4557-ae05-9ff2b0777eea\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.561300 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-config-data\") pod \"a110f476-90be-4557-ae05-9ff2b0777eea\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.561323 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a110f476-90be-4557-ae05-9ff2b0777eea-horizon-secret-key\") pod \"a110f476-90be-4557-ae05-9ff2b0777eea\" (UID: \"a110f476-90be-4557-ae05-9ff2b0777eea\") " Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.561657 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a110f476-90be-4557-ae05-9ff2b0777eea-logs" (OuterVolumeSpecName: "logs") pod "a110f476-90be-4557-ae05-9ff2b0777eea" (UID: "a110f476-90be-4557-ae05-9ff2b0777eea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.561949 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a110f476-90be-4557-ae05-9ff2b0777eea-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.566828 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a110f476-90be-4557-ae05-9ff2b0777eea-kube-api-access-h25fd" (OuterVolumeSpecName: "kube-api-access-h25fd") pod "a110f476-90be-4557-ae05-9ff2b0777eea" (UID: "a110f476-90be-4557-ae05-9ff2b0777eea"). InnerVolumeSpecName "kube-api-access-h25fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.567069 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a110f476-90be-4557-ae05-9ff2b0777eea-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a110f476-90be-4557-ae05-9ff2b0777eea" (UID: "a110f476-90be-4557-ae05-9ff2b0777eea"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.587505 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-scripts" (OuterVolumeSpecName: "scripts") pod "a110f476-90be-4557-ae05-9ff2b0777eea" (UID: "a110f476-90be-4557-ae05-9ff2b0777eea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.588088 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-config-data" (OuterVolumeSpecName: "config-data") pod "a110f476-90be-4557-ae05-9ff2b0777eea" (UID: "a110f476-90be-4557-ae05-9ff2b0777eea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.663830 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h25fd\" (UniqueName: \"kubernetes.io/projected/a110f476-90be-4557-ae05-9ff2b0777eea-kube-api-access-h25fd\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.663895 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.663905 4741 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a110f476-90be-4557-ae05-9ff2b0777eea-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.663916 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a110f476-90be-4557-ae05-9ff2b0777eea-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.797467 4741 generic.go:334] "Generic (PLEG): container finished" podID="a110f476-90be-4557-ae05-9ff2b0777eea" containerID="f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb" exitCode=137 Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.797507 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75b885c557-bbtjf" event={"ID":"a110f476-90be-4557-ae05-9ff2b0777eea","Type":"ContainerDied","Data":"f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb"} Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.797856 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75b885c557-bbtjf" event={"ID":"a110f476-90be-4557-ae05-9ff2b0777eea","Type":"ContainerDied","Data":"aebaf033a1d0139bb49b06210992dc609d54c399c2aff833b9394ee160f271d9"} Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.797875 4741 scope.go:117] "RemoveContainer" containerID="f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.797553 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75b885c557-bbtjf" Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.837454 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75b885c557-bbtjf"] Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.848372 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-75b885c557-bbtjf"] Sep 29 20:49:32 crc kubenswrapper[4741]: I0929 20:49:32.990002 4741 scope.go:117] "RemoveContainer" containerID="f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb" Sep 29 20:49:33 crc kubenswrapper[4741]: I0929 20:49:33.010492 4741 scope.go:117] "RemoveContainer" containerID="f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e" Sep 29 20:49:33 crc kubenswrapper[4741]: E0929 20:49:33.010899 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e\": container with ID starting with f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e not found: ID does not exist" containerID="f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e" Sep 29 20:49:33 crc kubenswrapper[4741]: I0929 20:49:33.010939 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e"} err="failed to get container status \"f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e\": rpc error: code = NotFound desc = could not find container \"f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e\": container with ID starting with f4f3fffb8fd99c78cc4a910f9d4d49acd22e9cec01d549e583738e3ba82ff27e not found: ID does not exist" Sep 29 20:49:33 crc kubenswrapper[4741]: I0929 20:49:33.010965 4741 scope.go:117] "RemoveContainer" containerID="f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb" Sep 29 20:49:33 crc kubenswrapper[4741]: E0929 20:49:33.011364 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb\": container with ID starting with f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb not found: ID does not exist" containerID="f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb" Sep 29 20:49:33 crc kubenswrapper[4741]: I0929 20:49:33.011422 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb"} err="failed to get container status \"f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb\": rpc error: code = NotFound desc = could not find container \"f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb\": container with ID starting with f4c303d6504de4df1c496cc20ecf6dbd9ee9aba2858bd08e3423b70909c8f4fb not found: ID does not exist" Sep 29 20:49:33 crc kubenswrapper[4741]: I0929 20:49:33.099143 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" path="/var/lib/kubelet/pods/a110f476-90be-4557-ae05-9ff2b0777eea/volumes" Sep 29 20:49:33 crc kubenswrapper[4741]: I0929 20:49:33.809249 4741 generic.go:334] "Generic (PLEG): container finished" podID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerID="d7a09c7862cd0e6994757d1f6b702b55a771f37dbd62fbade3b2a2e389e179b4" exitCode=0 Sep 29 20:49:33 crc kubenswrapper[4741]: I0929 20:49:33.809285 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ph5ql" event={"ID":"ff59d7a0-7c46-4595-9078-499ce72b7b7f","Type":"ContainerDied","Data":"d7a09c7862cd0e6994757d1f6b702b55a771f37dbd62fbade3b2a2e389e179b4"} Sep 29 20:49:34 crc kubenswrapper[4741]: I0929 20:49:34.829910 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ph5ql" event={"ID":"ff59d7a0-7c46-4595-9078-499ce72b7b7f","Type":"ContainerStarted","Data":"a04fee17458b12a49975fd5c917126ff85d267de84f51d1f800e73168990abf3"} Sep 29 20:49:34 crc kubenswrapper[4741]: I0929 20:49:34.866528 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ph5ql" podStartSLOduration=3.3029145939999998 podStartE2EDuration="5.866508102s" podCreationTimestamp="2025-09-29 20:49:29 +0000 UTC" firstStartedPulling="2025-09-29 20:49:31.790145465 +0000 UTC m=+6013.437934797" lastFinishedPulling="2025-09-29 20:49:34.353738973 +0000 UTC m=+6016.001528305" observedRunningTime="2025-09-29 20:49:34.858856266 +0000 UTC m=+6016.506645608" watchObservedRunningTime="2025-09-29 20:49:34.866508102 +0000 UTC m=+6016.514297444" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.499884 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lmkc9"] Sep 29 20:49:36 crc kubenswrapper[4741]: E0929 20:49:36.500721 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.500735 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" Sep 29 20:49:36 crc kubenswrapper[4741]: E0929 20:49:36.500752 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon-log" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.500758 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon-log" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.500963 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.500985 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a110f476-90be-4557-ae05-9ff2b0777eea" containerName="horizon-log" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.502524 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.524777 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lmkc9"] Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.642599 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tw5t\" (UniqueName: \"kubernetes.io/projected/2ce936bf-dd06-4da5-8e19-de25a71f47c0-kube-api-access-7tw5t\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.642987 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-catalog-content\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.643201 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-utilities\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.726024 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6f54844d7c-rrjpm"] Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.727739 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.745182 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6f54844d7c-rrjpm"] Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.745360 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tw5t\" (UniqueName: \"kubernetes.io/projected/2ce936bf-dd06-4da5-8e19-de25a71f47c0-kube-api-access-7tw5t\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.745443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-catalog-content\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.745490 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-utilities\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.746003 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-catalog-content\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.746028 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-utilities\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.770268 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tw5t\" (UniqueName: \"kubernetes.io/projected/2ce936bf-dd06-4da5-8e19-de25a71f47c0-kube-api-access-7tw5t\") pod \"certified-operators-lmkc9\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.869263 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.870103 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-scripts\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.870183 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-config-data\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.870237 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-logs\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.870313 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmh4k\" (UniqueName: \"kubernetes.io/projected/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-kube-api-access-bmh4k\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.870430 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-horizon-secret-key\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.972240 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-horizon-secret-key\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.972368 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-scripts\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.972427 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-config-data\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.972474 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-logs\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.972518 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmh4k\" (UniqueName: \"kubernetes.io/projected/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-kube-api-access-bmh4k\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.975173 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-scripts\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.975242 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-config-data\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.975529 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-logs\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.976878 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-horizon-secret-key\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:36 crc kubenswrapper[4741]: I0929 20:49:36.991880 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmh4k\" (UniqueName: \"kubernetes.io/projected/6195fc5e-a6c4-4a78-b45e-5c90ef096e3f-kube-api-access-bmh4k\") pod \"horizon-6f54844d7c-rrjpm\" (UID: \"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f\") " pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.050096 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.470809 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lmkc9"] Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.572533 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6f54844d7c-rrjpm"] Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.888342 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f54844d7c-rrjpm" event={"ID":"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f","Type":"ContainerStarted","Data":"63450b2a67cb3e68df2c922e3ebf6e432b3ce1e389c48ed47239c94ca9d38197"} Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.888980 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f54844d7c-rrjpm" event={"ID":"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f","Type":"ContainerStarted","Data":"53b083f635e87ed51dc61e5630a38e371d87a132af1de8d2433f11f35d928ab2"} Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.889792 4741 generic.go:334] "Generic (PLEG): container finished" podID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerID="5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8" exitCode=0 Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.889838 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmkc9" event={"ID":"2ce936bf-dd06-4da5-8e19-de25a71f47c0","Type":"ContainerDied","Data":"5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8"} Sep 29 20:49:37 crc kubenswrapper[4741]: I0929 20:49:37.889870 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmkc9" event={"ID":"2ce936bf-dd06-4da5-8e19-de25a71f47c0","Type":"ContainerStarted","Data":"458bd5f1efeffec6929042cd531aa89e82a5868e7d065d9f1d3b1c4d49347f77"} Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.315056 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-vz6r4"] Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.319216 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-vz6r4" Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.346482 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-vz6r4"] Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.511587 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57sm6\" (UniqueName: \"kubernetes.io/projected/943c6d41-b403-45de-891e-16c5421ab9da-kube-api-access-57sm6\") pod \"heat-db-create-vz6r4\" (UID: \"943c6d41-b403-45de-891e-16c5421ab9da\") " pod="openstack/heat-db-create-vz6r4" Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.614165 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57sm6\" (UniqueName: \"kubernetes.io/projected/943c6d41-b403-45de-891e-16c5421ab9da-kube-api-access-57sm6\") pod \"heat-db-create-vz6r4\" (UID: \"943c6d41-b403-45de-891e-16c5421ab9da\") " pod="openstack/heat-db-create-vz6r4" Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.633128 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57sm6\" (UniqueName: \"kubernetes.io/projected/943c6d41-b403-45de-891e-16c5421ab9da-kube-api-access-57sm6\") pod \"heat-db-create-vz6r4\" (UID: \"943c6d41-b403-45de-891e-16c5421ab9da\") " pod="openstack/heat-db-create-vz6r4" Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.641198 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-vz6r4" Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.905627 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f54844d7c-rrjpm" event={"ID":"6195fc5e-a6c4-4a78-b45e-5c90ef096e3f","Type":"ContainerStarted","Data":"2104e7f22db444244551e5ffd6528ab625c3c599c16f09f530cbdbcca2f46509"} Sep 29 20:49:38 crc kubenswrapper[4741]: I0929 20:49:38.931477 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6f54844d7c-rrjpm" podStartSLOduration=2.9314546679999998 podStartE2EDuration="2.931454668s" podCreationTimestamp="2025-09-29 20:49:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:49:38.9243526 +0000 UTC m=+6020.572141932" watchObservedRunningTime="2025-09-29 20:49:38.931454668 +0000 UTC m=+6020.579244010" Sep 29 20:49:39 crc kubenswrapper[4741]: I0929 20:49:39.177057 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-vz6r4"] Sep 29 20:49:39 crc kubenswrapper[4741]: W0929 20:49:39.186398 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod943c6d41_b403_45de_891e_16c5421ab9da.slice/crio-4362c270c7424eff808a37b95bc3ada4f04264f35e0e25777be2492bce2b58d2 WatchSource:0}: Error finding container 4362c270c7424eff808a37b95bc3ada4f04264f35e0e25777be2492bce2b58d2: Status 404 returned error can't find the container with id 4362c270c7424eff808a37b95bc3ada4f04264f35e0e25777be2492bce2b58d2 Sep 29 20:49:39 crc kubenswrapper[4741]: I0929 20:49:39.919038 4741 generic.go:334] "Generic (PLEG): container finished" podID="943c6d41-b403-45de-891e-16c5421ab9da" containerID="35338d1b61e585850d495e402eb13ffb9f6a5dec35f6a6c2a4c281db31552a5a" exitCode=0 Sep 29 20:49:39 crc kubenswrapper[4741]: I0929 20:49:39.919090 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-vz6r4" event={"ID":"943c6d41-b403-45de-891e-16c5421ab9da","Type":"ContainerDied","Data":"35338d1b61e585850d495e402eb13ffb9f6a5dec35f6a6c2a4c281db31552a5a"} Sep 29 20:49:39 crc kubenswrapper[4741]: I0929 20:49:39.919136 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-vz6r4" event={"ID":"943c6d41-b403-45de-891e-16c5421ab9da","Type":"ContainerStarted","Data":"4362c270c7424eff808a37b95bc3ada4f04264f35e0e25777be2492bce2b58d2"} Sep 29 20:49:39 crc kubenswrapper[4741]: I0929 20:49:39.922785 4741 generic.go:334] "Generic (PLEG): container finished" podID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerID="7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0" exitCode=0 Sep 29 20:49:39 crc kubenswrapper[4741]: I0929 20:49:39.922910 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmkc9" event={"ID":"2ce936bf-dd06-4da5-8e19-de25a71f47c0","Type":"ContainerDied","Data":"7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0"} Sep 29 20:49:40 crc kubenswrapper[4741]: I0929 20:49:40.242111 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:40 crc kubenswrapper[4741]: I0929 20:49:40.242452 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:40 crc kubenswrapper[4741]: I0929 20:49:40.294719 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:40 crc kubenswrapper[4741]: I0929 20:49:40.936839 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmkc9" event={"ID":"2ce936bf-dd06-4da5-8e19-de25a71f47c0","Type":"ContainerStarted","Data":"1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9"} Sep 29 20:49:40 crc kubenswrapper[4741]: I0929 20:49:40.969054 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lmkc9" podStartSLOduration=2.400861505 podStartE2EDuration="4.96903322s" podCreationTimestamp="2025-09-29 20:49:36 +0000 UTC" firstStartedPulling="2025-09-29 20:49:37.891479425 +0000 UTC m=+6019.539268757" lastFinishedPulling="2025-09-29 20:49:40.45965114 +0000 UTC m=+6022.107440472" observedRunningTime="2025-09-29 20:49:40.963335996 +0000 UTC m=+6022.611125328" watchObservedRunningTime="2025-09-29 20:49:40.96903322 +0000 UTC m=+6022.616822552" Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.002754 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.086219 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:49:41 crc kubenswrapper[4741]: E0929 20:49:41.086568 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.346598 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-vz6r4" Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.483353 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57sm6\" (UniqueName: \"kubernetes.io/projected/943c6d41-b403-45de-891e-16c5421ab9da-kube-api-access-57sm6\") pod \"943c6d41-b403-45de-891e-16c5421ab9da\" (UID: \"943c6d41-b403-45de-891e-16c5421ab9da\") " Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.500686 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/943c6d41-b403-45de-891e-16c5421ab9da-kube-api-access-57sm6" (OuterVolumeSpecName: "kube-api-access-57sm6") pod "943c6d41-b403-45de-891e-16c5421ab9da" (UID: "943c6d41-b403-45de-891e-16c5421ab9da"). InnerVolumeSpecName "kube-api-access-57sm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.586050 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57sm6\" (UniqueName: \"kubernetes.io/projected/943c6d41-b403-45de-891e-16c5421ab9da-kube-api-access-57sm6\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.946510 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-vz6r4" event={"ID":"943c6d41-b403-45de-891e-16c5421ab9da","Type":"ContainerDied","Data":"4362c270c7424eff808a37b95bc3ada4f04264f35e0e25777be2492bce2b58d2"} Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.947509 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4362c270c7424eff808a37b95bc3ada4f04264f35e0e25777be2492bce2b58d2" Sep 29 20:49:41 crc kubenswrapper[4741]: I0929 20:49:41.946584 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-vz6r4" Sep 29 20:49:45 crc kubenswrapper[4741]: I0929 20:49:45.666714 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ph5ql"] Sep 29 20:49:45 crc kubenswrapper[4741]: I0929 20:49:45.667426 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ph5ql" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="registry-server" containerID="cri-o://a04fee17458b12a49975fd5c917126ff85d267de84f51d1f800e73168990abf3" gracePeriod=2 Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:45.986192 4741 generic.go:334] "Generic (PLEG): container finished" podID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerID="a04fee17458b12a49975fd5c917126ff85d267de84f51d1f800e73168990abf3" exitCode=0 Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:45.986262 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ph5ql" event={"ID":"ff59d7a0-7c46-4595-9078-499ce72b7b7f","Type":"ContainerDied","Data":"a04fee17458b12a49975fd5c917126ff85d267de84f51d1f800e73168990abf3"} Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.215043 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.374891 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-catalog-content\") pod \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.374960 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-utilities\") pod \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.375033 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4q76\" (UniqueName: \"kubernetes.io/projected/ff59d7a0-7c46-4595-9078-499ce72b7b7f-kube-api-access-n4q76\") pod \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\" (UID: \"ff59d7a0-7c46-4595-9078-499ce72b7b7f\") " Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.375811 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-utilities" (OuterVolumeSpecName: "utilities") pod "ff59d7a0-7c46-4595-9078-499ce72b7b7f" (UID: "ff59d7a0-7c46-4595-9078-499ce72b7b7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.387012 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff59d7a0-7c46-4595-9078-499ce72b7b7f" (UID: "ff59d7a0-7c46-4595-9078-499ce72b7b7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.387589 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff59d7a0-7c46-4595-9078-499ce72b7b7f-kube-api-access-n4q76" (OuterVolumeSpecName: "kube-api-access-n4q76") pod "ff59d7a0-7c46-4595-9078-499ce72b7b7f" (UID: "ff59d7a0-7c46-4595-9078-499ce72b7b7f"). InnerVolumeSpecName "kube-api-access-n4q76". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.477615 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.477645 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff59d7a0-7c46-4595-9078-499ce72b7b7f-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.477661 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4q76\" (UniqueName: \"kubernetes.io/projected/ff59d7a0-7c46-4595-9078-499ce72b7b7f-kube-api-access-n4q76\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.869446 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.870666 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.938817 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:46 crc kubenswrapper[4741]: I0929 20:49:46.997920 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ph5ql" Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.009042 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ph5ql" event={"ID":"ff59d7a0-7c46-4595-9078-499ce72b7b7f","Type":"ContainerDied","Data":"d1ffeb121c6ee79f2877f4c2b551d29c72a033acd9f6a2ac9f1e5f13b9d0b054"} Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.009181 4741 scope.go:117] "RemoveContainer" containerID="a04fee17458b12a49975fd5c917126ff85d267de84f51d1f800e73168990abf3" Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.036252 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ph5ql"] Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.037734 4741 scope.go:117] "RemoveContainer" containerID="d7a09c7862cd0e6994757d1f6b702b55a771f37dbd62fbade3b2a2e389e179b4" Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.045945 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ph5ql"] Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.054898 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.055112 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.063608 4741 scope.go:117] "RemoveContainer" containerID="3d3b48e4ccc8051edcbf97afbf91c86def81c96532f0c9fe5e3a1987e130076b" Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.071186 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:47 crc kubenswrapper[4741]: I0929 20:49:47.101262 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" path="/var/lib/kubelet/pods/ff59d7a0-7c46-4595-9078-499ce72b7b7f/volumes" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.069704 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lmkc9"] Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.380884 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-a79d-account-create-jm4qm"] Sep 29 20:49:48 crc kubenswrapper[4741]: E0929 20:49:48.381755 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="extract-content" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.381798 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="extract-content" Sep 29 20:49:48 crc kubenswrapper[4741]: E0929 20:49:48.381836 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="943c6d41-b403-45de-891e-16c5421ab9da" containerName="mariadb-database-create" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.381856 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="943c6d41-b403-45de-891e-16c5421ab9da" containerName="mariadb-database-create" Sep 29 20:49:48 crc kubenswrapper[4741]: E0929 20:49:48.381895 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="extract-utilities" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.381912 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="extract-utilities" Sep 29 20:49:48 crc kubenswrapper[4741]: E0929 20:49:48.381953 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="registry-server" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.381969 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="registry-server" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.382449 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff59d7a0-7c46-4595-9078-499ce72b7b7f" containerName="registry-server" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.382485 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="943c6d41-b403-45de-891e-16c5421ab9da" containerName="mariadb-database-create" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.385937 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a79d-account-create-jm4qm" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.389365 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a79d-account-create-jm4qm"] Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.390818 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.521245 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/aee1d090-e1c4-47ce-8bc9-03fa600db9fa-kube-api-access-2rddx\") pod \"heat-a79d-account-create-jm4qm\" (UID: \"aee1d090-e1c4-47ce-8bc9-03fa600db9fa\") " pod="openstack/heat-a79d-account-create-jm4qm" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.623416 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/aee1d090-e1c4-47ce-8bc9-03fa600db9fa-kube-api-access-2rddx\") pod \"heat-a79d-account-create-jm4qm\" (UID: \"aee1d090-e1c4-47ce-8bc9-03fa600db9fa\") " pod="openstack/heat-a79d-account-create-jm4qm" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.640773 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/aee1d090-e1c4-47ce-8bc9-03fa600db9fa-kube-api-access-2rddx\") pod \"heat-a79d-account-create-jm4qm\" (UID: \"aee1d090-e1c4-47ce-8bc9-03fa600db9fa\") " pod="openstack/heat-a79d-account-create-jm4qm" Sep 29 20:49:48 crc kubenswrapper[4741]: I0929 20:49:48.714585 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a79d-account-create-jm4qm" Sep 29 20:49:49 crc kubenswrapper[4741]: I0929 20:49:49.176200 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a79d-account-create-jm4qm"] Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.020914 4741 generic.go:334] "Generic (PLEG): container finished" podID="aee1d090-e1c4-47ce-8bc9-03fa600db9fa" containerID="26686b3dbab6234575f00d623fbe300fe9c2d7e583e9ba940295cf316fc259c3" exitCode=0 Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.021018 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a79d-account-create-jm4qm" event={"ID":"aee1d090-e1c4-47ce-8bc9-03fa600db9fa","Type":"ContainerDied","Data":"26686b3dbab6234575f00d623fbe300fe9c2d7e583e9ba940295cf316fc259c3"} Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.021280 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a79d-account-create-jm4qm" event={"ID":"aee1d090-e1c4-47ce-8bc9-03fa600db9fa","Type":"ContainerStarted","Data":"cd75a9c8065a4d69b126ec184a9cb548f60ae54eccf569907a83cb33e4dfacfc"} Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.021505 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lmkc9" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="registry-server" containerID="cri-o://1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9" gracePeriod=2 Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.498289 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.663663 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-catalog-content\") pod \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.663895 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tw5t\" (UniqueName: \"kubernetes.io/projected/2ce936bf-dd06-4da5-8e19-de25a71f47c0-kube-api-access-7tw5t\") pod \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.663979 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-utilities\") pod \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\" (UID: \"2ce936bf-dd06-4da5-8e19-de25a71f47c0\") " Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.665702 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-utilities" (OuterVolumeSpecName: "utilities") pod "2ce936bf-dd06-4da5-8e19-de25a71f47c0" (UID: "2ce936bf-dd06-4da5-8e19-de25a71f47c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.681354 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ce936bf-dd06-4da5-8e19-de25a71f47c0-kube-api-access-7tw5t" (OuterVolumeSpecName: "kube-api-access-7tw5t") pod "2ce936bf-dd06-4da5-8e19-de25a71f47c0" (UID: "2ce936bf-dd06-4da5-8e19-de25a71f47c0"). InnerVolumeSpecName "kube-api-access-7tw5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.730110 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ce936bf-dd06-4da5-8e19-de25a71f47c0" (UID: "2ce936bf-dd06-4da5-8e19-de25a71f47c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.767718 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.767767 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tw5t\" (UniqueName: \"kubernetes.io/projected/2ce936bf-dd06-4da5-8e19-de25a71f47c0-kube-api-access-7tw5t\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:50 crc kubenswrapper[4741]: I0929 20:49:50.767787 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce936bf-dd06-4da5-8e19-de25a71f47c0-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.039779 4741 generic.go:334] "Generic (PLEG): container finished" podID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerID="1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9" exitCode=0 Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.039859 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmkc9" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.039919 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmkc9" event={"ID":"2ce936bf-dd06-4da5-8e19-de25a71f47c0","Type":"ContainerDied","Data":"1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9"} Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.045641 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmkc9" event={"ID":"2ce936bf-dd06-4da5-8e19-de25a71f47c0","Type":"ContainerDied","Data":"458bd5f1efeffec6929042cd531aa89e82a5868e7d065d9f1d3b1c4d49347f77"} Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.045677 4741 scope.go:117] "RemoveContainer" containerID="1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.088657 4741 scope.go:117] "RemoveContainer" containerID="7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.107531 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lmkc9"] Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.117619 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lmkc9"] Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.138579 4741 scope.go:117] "RemoveContainer" containerID="5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.189121 4741 scope.go:117] "RemoveContainer" containerID="1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9" Sep 29 20:49:51 crc kubenswrapper[4741]: E0929 20:49:51.189709 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9\": container with ID starting with 1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9 not found: ID does not exist" containerID="1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.189768 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9"} err="failed to get container status \"1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9\": rpc error: code = NotFound desc = could not find container \"1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9\": container with ID starting with 1555c5e1598053b4a9d4a8664cd02c70fb5d548176f92ced79a6562079ba50a9 not found: ID does not exist" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.189796 4741 scope.go:117] "RemoveContainer" containerID="7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0" Sep 29 20:49:51 crc kubenswrapper[4741]: E0929 20:49:51.190236 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0\": container with ID starting with 7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0 not found: ID does not exist" containerID="7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.190293 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0"} err="failed to get container status \"7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0\": rpc error: code = NotFound desc = could not find container \"7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0\": container with ID starting with 7de345910c84d9c7846379ee599def21d5e23788641da433e5342e935480b6a0 not found: ID does not exist" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.190329 4741 scope.go:117] "RemoveContainer" containerID="5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8" Sep 29 20:49:51 crc kubenswrapper[4741]: E0929 20:49:51.190810 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8\": container with ID starting with 5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8 not found: ID does not exist" containerID="5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.190888 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8"} err="failed to get container status \"5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8\": rpc error: code = NotFound desc = could not find container \"5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8\": container with ID starting with 5c0964a734bac085e0a727cffbf3e6668e3286509c6c209af38a3eda99571fb8 not found: ID does not exist" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.535491 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a79d-account-create-jm4qm" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.599375 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/aee1d090-e1c4-47ce-8bc9-03fa600db9fa-kube-api-access-2rddx\") pod \"aee1d090-e1c4-47ce-8bc9-03fa600db9fa\" (UID: \"aee1d090-e1c4-47ce-8bc9-03fa600db9fa\") " Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.619606 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aee1d090-e1c4-47ce-8bc9-03fa600db9fa-kube-api-access-2rddx" (OuterVolumeSpecName: "kube-api-access-2rddx") pod "aee1d090-e1c4-47ce-8bc9-03fa600db9fa" (UID: "aee1d090-e1c4-47ce-8bc9-03fa600db9fa"). InnerVolumeSpecName "kube-api-access-2rddx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:49:51 crc kubenswrapper[4741]: I0929 20:49:51.703510 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/aee1d090-e1c4-47ce-8bc9-03fa600db9fa-kube-api-access-2rddx\") on node \"crc\" DevicePath \"\"" Sep 29 20:49:52 crc kubenswrapper[4741]: I0929 20:49:52.055695 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a79d-account-create-jm4qm" Sep 29 20:49:52 crc kubenswrapper[4741]: I0929 20:49:52.056081 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a79d-account-create-jm4qm" event={"ID":"aee1d090-e1c4-47ce-8bc9-03fa600db9fa","Type":"ContainerDied","Data":"cd75a9c8065a4d69b126ec184a9cb548f60ae54eccf569907a83cb33e4dfacfc"} Sep 29 20:49:52 crc kubenswrapper[4741]: I0929 20:49:52.056137 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd75a9c8065a4d69b126ec184a9cb548f60ae54eccf569907a83cb33e4dfacfc" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.105258 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" path="/var/lib/kubelet/pods/2ce936bf-dd06-4da5-8e19-de25a71f47c0/volumes" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.453289 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-d5dz9"] Sep 29 20:49:53 crc kubenswrapper[4741]: E0929 20:49:53.454054 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="extract-utilities" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.454073 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="extract-utilities" Sep 29 20:49:53 crc kubenswrapper[4741]: E0929 20:49:53.454090 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="extract-content" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.454098 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="extract-content" Sep 29 20:49:53 crc kubenswrapper[4741]: E0929 20:49:53.454109 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee1d090-e1c4-47ce-8bc9-03fa600db9fa" containerName="mariadb-account-create" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.454115 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee1d090-e1c4-47ce-8bc9-03fa600db9fa" containerName="mariadb-account-create" Sep 29 20:49:53 crc kubenswrapper[4741]: E0929 20:49:53.454141 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="registry-server" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.454146 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="registry-server" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.454315 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee1d090-e1c4-47ce-8bc9-03fa600db9fa" containerName="mariadb-account-create" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.454335 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ce936bf-dd06-4da5-8e19-de25a71f47c0" containerName="registry-server" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.455028 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.457695 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-frfj4" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.457912 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.474059 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-d5dz9"] Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.547678 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-config-data\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.547934 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d8hc\" (UniqueName: \"kubernetes.io/projected/302998d5-a7fc-458f-93e4-6e1893537638-kube-api-access-9d8hc\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.547970 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-combined-ca-bundle\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.650231 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d8hc\" (UniqueName: \"kubernetes.io/projected/302998d5-a7fc-458f-93e4-6e1893537638-kube-api-access-9d8hc\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.650366 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-combined-ca-bundle\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.650593 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-config-data\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.656500 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-combined-ca-bundle\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.673674 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-config-data\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.678134 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d8hc\" (UniqueName: \"kubernetes.io/projected/302998d5-a7fc-458f-93e4-6e1893537638-kube-api-access-9d8hc\") pod \"heat-db-sync-d5dz9\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:53 crc kubenswrapper[4741]: I0929 20:49:53.807409 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-d5dz9" Sep 29 20:49:54 crc kubenswrapper[4741]: W0929 20:49:54.291906 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302998d5_a7fc_458f_93e4_6e1893537638.slice/crio-ad6c228ba3a603c4d3728c85bbf53dac109fed4094ceaf714046655cbb0f3986 WatchSource:0}: Error finding container ad6c228ba3a603c4d3728c85bbf53dac109fed4094ceaf714046655cbb0f3986: Status 404 returned error can't find the container with id ad6c228ba3a603c4d3728c85bbf53dac109fed4094ceaf714046655cbb0f3986 Sep 29 20:49:54 crc kubenswrapper[4741]: I0929 20:49:54.292980 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-d5dz9"] Sep 29 20:49:55 crc kubenswrapper[4741]: I0929 20:49:55.101669 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-d5dz9" event={"ID":"302998d5-a7fc-458f-93e4-6e1893537638","Type":"ContainerStarted","Data":"ad6c228ba3a603c4d3728c85bbf53dac109fed4094ceaf714046655cbb0f3986"} Sep 29 20:49:56 crc kubenswrapper[4741]: I0929 20:49:56.086343 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:49:56 crc kubenswrapper[4741]: E0929 20:49:56.086929 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:49:57 crc kubenswrapper[4741]: I0929 20:49:57.047956 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-4n6h7"] Sep 29 20:49:57 crc kubenswrapper[4741]: I0929 20:49:57.062330 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-4n6h7"] Sep 29 20:49:57 crc kubenswrapper[4741]: I0929 20:49:57.098977 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="922b6d69-6d18-4a0c-b81a-e2aa6d103c2f" path="/var/lib/kubelet/pods/922b6d69-6d18-4a0c-b81a-e2aa6d103c2f/volumes" Sep 29 20:49:58 crc kubenswrapper[4741]: I0929 20:49:58.855033 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:50:00 crc kubenswrapper[4741]: I0929 20:50:00.549103 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6f54844d7c-rrjpm" Sep 29 20:50:00 crc kubenswrapper[4741]: I0929 20:50:00.610022 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6468bdd6cc-7gxj7"] Sep 29 20:50:00 crc kubenswrapper[4741]: I0929 20:50:00.610247 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6468bdd6cc-7gxj7" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon-log" containerID="cri-o://d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58" gracePeriod=30 Sep 29 20:50:00 crc kubenswrapper[4741]: I0929 20:50:00.610341 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6468bdd6cc-7gxj7" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" containerID="cri-o://30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4" gracePeriod=30 Sep 29 20:50:02 crc kubenswrapper[4741]: I0929 20:50:02.169791 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-d5dz9" event={"ID":"302998d5-a7fc-458f-93e4-6e1893537638","Type":"ContainerStarted","Data":"7c37a22716f28cff2a47266b8c405cc1ae739bfe4d51105630a8156a6be79cab"} Sep 29 20:50:02 crc kubenswrapper[4741]: I0929 20:50:02.195700 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-d5dz9" podStartSLOduration=2.346354581 podStartE2EDuration="9.195683473s" podCreationTimestamp="2025-09-29 20:49:53 +0000 UTC" firstStartedPulling="2025-09-29 20:49:54.294100055 +0000 UTC m=+6035.941889387" lastFinishedPulling="2025-09-29 20:50:01.143428947 +0000 UTC m=+6042.791218279" observedRunningTime="2025-09-29 20:50:02.189996534 +0000 UTC m=+6043.837785876" watchObservedRunningTime="2025-09-29 20:50:02.195683473 +0000 UTC m=+6043.843472805" Sep 29 20:50:03 crc kubenswrapper[4741]: I0929 20:50:03.186911 4741 generic.go:334] "Generic (PLEG): container finished" podID="302998d5-a7fc-458f-93e4-6e1893537638" containerID="7c37a22716f28cff2a47266b8c405cc1ae739bfe4d51105630a8156a6be79cab" exitCode=0 Sep 29 20:50:03 crc kubenswrapper[4741]: I0929 20:50:03.187266 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-d5dz9" event={"ID":"302998d5-a7fc-458f-93e4-6e1893537638","Type":"ContainerDied","Data":"7c37a22716f28cff2a47266b8c405cc1ae739bfe4d51105630a8156a6be79cab"} Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.202950 4741 generic.go:334] "Generic (PLEG): container finished" podID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerID="30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4" exitCode=0 Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.203009 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6468bdd6cc-7gxj7" event={"ID":"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15","Type":"ContainerDied","Data":"30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4"} Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.560438 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-d5dz9" Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.678935 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-combined-ca-bundle\") pod \"302998d5-a7fc-458f-93e4-6e1893537638\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.678994 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-config-data\") pod \"302998d5-a7fc-458f-93e4-6e1893537638\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.679024 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d8hc\" (UniqueName: \"kubernetes.io/projected/302998d5-a7fc-458f-93e4-6e1893537638-kube-api-access-9d8hc\") pod \"302998d5-a7fc-458f-93e4-6e1893537638\" (UID: \"302998d5-a7fc-458f-93e4-6e1893537638\") " Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.683686 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302998d5-a7fc-458f-93e4-6e1893537638-kube-api-access-9d8hc" (OuterVolumeSpecName: "kube-api-access-9d8hc") pod "302998d5-a7fc-458f-93e4-6e1893537638" (UID: "302998d5-a7fc-458f-93e4-6e1893537638"). InnerVolumeSpecName "kube-api-access-9d8hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.706304 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "302998d5-a7fc-458f-93e4-6e1893537638" (UID: "302998d5-a7fc-458f-93e4-6e1893537638"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.761627 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-config-data" (OuterVolumeSpecName: "config-data") pod "302998d5-a7fc-458f-93e4-6e1893537638" (UID: "302998d5-a7fc-458f-93e4-6e1893537638"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.781536 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.781565 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302998d5-a7fc-458f-93e4-6e1893537638-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:04 crc kubenswrapper[4741]: I0929 20:50:04.781575 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d8hc\" (UniqueName: \"kubernetes.io/projected/302998d5-a7fc-458f-93e4-6e1893537638-kube-api-access-9d8hc\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:05 crc kubenswrapper[4741]: I0929 20:50:05.223731 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-d5dz9" event={"ID":"302998d5-a7fc-458f-93e4-6e1893537638","Type":"ContainerDied","Data":"ad6c228ba3a603c4d3728c85bbf53dac109fed4094ceaf714046655cbb0f3986"} Sep 29 20:50:05 crc kubenswrapper[4741]: I0929 20:50:05.224044 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad6c228ba3a603c4d3728c85bbf53dac109fed4094ceaf714046655cbb0f3986" Sep 29 20:50:05 crc kubenswrapper[4741]: I0929 20:50:05.223767 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-d5dz9" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.395634 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7695685f8c-jtmfd"] Sep 29 20:50:06 crc kubenswrapper[4741]: E0929 20:50:06.396994 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302998d5-a7fc-458f-93e4-6e1893537638" containerName="heat-db-sync" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.397016 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="302998d5-a7fc-458f-93e4-6e1893537638" containerName="heat-db-sync" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.397252 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="302998d5-a7fc-458f-93e4-6e1893537638" containerName="heat-db-sync" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.398108 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.400324 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.400718 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-frfj4" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.403452 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.448764 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7695685f8c-jtmfd"] Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.516292 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-combined-ca-bundle\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.516344 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt6c2\" (UniqueName: \"kubernetes.io/projected/65abba08-dad1-4b41-a663-4d56c2b152e4-kube-api-access-wt6c2\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.516383 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-config-data\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.516496 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-config-data-custom\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.558967 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6df6b5f465-mp74v"] Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.561329 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.568743 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.594824 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-67f555958f-bpf7z"] Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.596064 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.598055 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.607584 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6df6b5f465-mp74v"] Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.617916 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt6c2\" (UniqueName: \"kubernetes.io/projected/65abba08-dad1-4b41-a663-4d56c2b152e4-kube-api-access-wt6c2\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.617980 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-config-data\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.618075 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db94g\" (UniqueName: \"kubernetes.io/projected/c94210e6-f88d-472f-9048-5511f940db0c-kube-api-access-db94g\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.618115 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-config-data-custom\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.618137 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-config-data-custom\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.618175 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-combined-ca-bundle\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.618198 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-config-data\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.619186 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-combined-ca-bundle\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.624850 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-config-data\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.625089 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-combined-ca-bundle\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.626591 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65abba08-dad1-4b41-a663-4d56c2b152e4-config-data-custom\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.627171 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-67f555958f-bpf7z"] Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.634267 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt6c2\" (UniqueName: \"kubernetes.io/projected/65abba08-dad1-4b41-a663-4d56c2b152e4-kube-api-access-wt6c2\") pod \"heat-engine-7695685f8c-jtmfd\" (UID: \"65abba08-dad1-4b41-a663-4d56c2b152e4\") " pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721297 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-config-data\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721347 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db94g\" (UniqueName: \"kubernetes.io/projected/c94210e6-f88d-472f-9048-5511f940db0c-kube-api-access-db94g\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721405 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-config-data-custom\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721445 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-combined-ca-bundle\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721469 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-config-data\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721508 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-config-data-custom\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721549 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hznpr\" (UniqueName: \"kubernetes.io/projected/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-kube-api-access-hznpr\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.721604 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-combined-ca-bundle\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.728260 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-combined-ca-bundle\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.728571 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-config-data\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.728576 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c94210e6-f88d-472f-9048-5511f940db0c-config-data-custom\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.733860 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.742759 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db94g\" (UniqueName: \"kubernetes.io/projected/c94210e6-f88d-472f-9048-5511f940db0c-kube-api-access-db94g\") pod \"heat-cfnapi-6df6b5f465-mp74v\" (UID: \"c94210e6-f88d-472f-9048-5511f940db0c\") " pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.823509 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hznpr\" (UniqueName: \"kubernetes.io/projected/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-kube-api-access-hznpr\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.823616 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-combined-ca-bundle\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.823668 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-config-data\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.823798 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-config-data-custom\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.829215 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-combined-ca-bundle\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.830572 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-config-data\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.838444 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-config-data-custom\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.856293 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hznpr\" (UniqueName: \"kubernetes.io/projected/c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407-kube-api-access-hznpr\") pod \"heat-api-67f555958f-bpf7z\" (UID: \"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407\") " pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.896818 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:06 crc kubenswrapper[4741]: I0929 20:50:06.913627 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:07 crc kubenswrapper[4741]: I0929 20:50:07.049527 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a214-account-create-jlwc9"] Sep 29 20:50:07 crc kubenswrapper[4741]: I0929 20:50:07.061717 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a214-account-create-jlwc9"] Sep 29 20:50:07 crc kubenswrapper[4741]: I0929 20:50:07.103311 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5007e794-725b-48f2-916b-8e412e4780fc" path="/var/lib/kubelet/pods/5007e794-725b-48f2-916b-8e412e4780fc/volumes" Sep 29 20:50:07 crc kubenswrapper[4741]: I0929 20:50:07.211136 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7695685f8c-jtmfd"] Sep 29 20:50:07 crc kubenswrapper[4741]: I0929 20:50:07.248122 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7695685f8c-jtmfd" event={"ID":"65abba08-dad1-4b41-a663-4d56c2b152e4","Type":"ContainerStarted","Data":"83c4b0f9b5cf6402bbc0514ce3d1c5cb2629a8d922f0ca27ff78b20ece57eea3"} Sep 29 20:50:07 crc kubenswrapper[4741]: W0929 20:50:07.437064 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8d784d6_ef0b_4f31_a1c6_cbc08cbd2407.slice/crio-bc41fd5e70718cf809d5ad830269b03c652f9dc3e787d9c894a497e4438d577e WatchSource:0}: Error finding container bc41fd5e70718cf809d5ad830269b03c652f9dc3e787d9c894a497e4438d577e: Status 404 returned error can't find the container with id bc41fd5e70718cf809d5ad830269b03c652f9dc3e787d9c894a497e4438d577e Sep 29 20:50:07 crc kubenswrapper[4741]: I0929 20:50:07.444589 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-67f555958f-bpf7z"] Sep 29 20:50:07 crc kubenswrapper[4741]: I0929 20:50:07.452931 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6df6b5f465-mp74v"] Sep 29 20:50:08 crc kubenswrapper[4741]: I0929 20:50:08.262580 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6df6b5f465-mp74v" event={"ID":"c94210e6-f88d-472f-9048-5511f940db0c","Type":"ContainerStarted","Data":"0c206a644fc2eb10e329b8bea68c7b1ce7ead10310484246ec7a0ee3ea40e61b"} Sep 29 20:50:08 crc kubenswrapper[4741]: I0929 20:50:08.265920 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7695685f8c-jtmfd" event={"ID":"65abba08-dad1-4b41-a663-4d56c2b152e4","Type":"ContainerStarted","Data":"8c019cf1a8d189f2a9fd1e8f23a158d93e550cb5388c3a13a981d8e3d23ed330"} Sep 29 20:50:08 crc kubenswrapper[4741]: I0929 20:50:08.266883 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:08 crc kubenswrapper[4741]: I0929 20:50:08.271807 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-67f555958f-bpf7z" event={"ID":"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407","Type":"ContainerStarted","Data":"bc41fd5e70718cf809d5ad830269b03c652f9dc3e787d9c894a497e4438d577e"} Sep 29 20:50:08 crc kubenswrapper[4741]: I0929 20:50:08.288476 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7695685f8c-jtmfd" podStartSLOduration=2.28845428 podStartE2EDuration="2.28845428s" podCreationTimestamp="2025-09-29 20:50:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:50:08.283168364 +0000 UTC m=+6049.930957696" watchObservedRunningTime="2025-09-29 20:50:08.28845428 +0000 UTC m=+6049.936243612" Sep 29 20:50:08 crc kubenswrapper[4741]: I0929 20:50:08.463941 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6468bdd6cc-7gxj7" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Sep 29 20:50:10 crc kubenswrapper[4741]: I0929 20:50:10.303070 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-67f555958f-bpf7z" event={"ID":"c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407","Type":"ContainerStarted","Data":"6135a45bf8fb0ea04fcba8b5b58a95f296dd307b311341016ad64eb63a345c70"} Sep 29 20:50:10 crc kubenswrapper[4741]: I0929 20:50:10.304250 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:10 crc kubenswrapper[4741]: I0929 20:50:10.306105 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6df6b5f465-mp74v" event={"ID":"c94210e6-f88d-472f-9048-5511f940db0c","Type":"ContainerStarted","Data":"4a40b04ed8f8653422921d81068bdf331fff029de27f7c4c644ed0dba731683c"} Sep 29 20:50:10 crc kubenswrapper[4741]: I0929 20:50:10.306283 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:10 crc kubenswrapper[4741]: I0929 20:50:10.321250 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-67f555958f-bpf7z" podStartSLOduration=2.320728607 podStartE2EDuration="4.321234794s" podCreationTimestamp="2025-09-29 20:50:06 +0000 UTC" firstStartedPulling="2025-09-29 20:50:07.444345079 +0000 UTC m=+6049.092134411" lastFinishedPulling="2025-09-29 20:50:09.444851266 +0000 UTC m=+6051.092640598" observedRunningTime="2025-09-29 20:50:10.319923672 +0000 UTC m=+6051.967713014" watchObservedRunningTime="2025-09-29 20:50:10.321234794 +0000 UTC m=+6051.969024126" Sep 29 20:50:10 crc kubenswrapper[4741]: I0929 20:50:10.338013 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6df6b5f465-mp74v" podStartSLOduration=2.33671361 podStartE2EDuration="4.337996182s" podCreationTimestamp="2025-09-29 20:50:06 +0000 UTC" firstStartedPulling="2025-09-29 20:50:07.445205925 +0000 UTC m=+6049.092995257" lastFinishedPulling="2025-09-29 20:50:09.446488497 +0000 UTC m=+6051.094277829" observedRunningTime="2025-09-29 20:50:10.336563066 +0000 UTC m=+6051.984352398" watchObservedRunningTime="2025-09-29 20:50:10.337996182 +0000 UTC m=+6051.985785514" Sep 29 20:50:11 crc kubenswrapper[4741]: I0929 20:50:11.086108 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:50:11 crc kubenswrapper[4741]: E0929 20:50:11.086644 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:50:14 crc kubenswrapper[4741]: I0929 20:50:14.031096 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nwzl6"] Sep 29 20:50:14 crc kubenswrapper[4741]: I0929 20:50:14.040993 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nwzl6"] Sep 29 20:50:15 crc kubenswrapper[4741]: I0929 20:50:15.109028 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a716d83-536e-41ad-b8df-b55cfbfac7a1" path="/var/lib/kubelet/pods/7a716d83-536e-41ad-b8df-b55cfbfac7a1/volumes" Sep 29 20:50:18 crc kubenswrapper[4741]: I0929 20:50:18.155784 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-6df6b5f465-mp74v" Sep 29 20:50:18 crc kubenswrapper[4741]: I0929 20:50:18.301425 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-67f555958f-bpf7z" Sep 29 20:50:18 crc kubenswrapper[4741]: I0929 20:50:18.464576 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6468bdd6cc-7gxj7" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Sep 29 20:50:23 crc kubenswrapper[4741]: I0929 20:50:23.086355 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:50:23 crc kubenswrapper[4741]: E0929 20:50:23.087230 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:50:26 crc kubenswrapper[4741]: I0929 20:50:26.763738 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7695685f8c-jtmfd" Sep 29 20:50:28 crc kubenswrapper[4741]: I0929 20:50:28.463608 4741 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6468bdd6cc-7gxj7" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Sep 29 20:50:28 crc kubenswrapper[4741]: I0929 20:50:28.464048 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:50:28 crc kubenswrapper[4741]: I0929 20:50:28.681501 4741 scope.go:117] "RemoveContainer" containerID="e62edaf61dadabb033b60c92a8ef4dcbc4e9791d8b47a597eeb00714c7a2328a" Sep 29 20:50:28 crc kubenswrapper[4741]: I0929 20:50:28.720065 4741 scope.go:117] "RemoveContainer" containerID="26e8685f0b49214da518ce90beb6aa2904ce8226b2bb7c38cd7954a956920cef" Sep 29 20:50:28 crc kubenswrapper[4741]: I0929 20:50:28.772884 4741 scope.go:117] "RemoveContainer" containerID="fef575ca088dbdbd8d09178b957915cfb88901de31de6029240d451ec60c97a1" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.148105 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.237864 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-horizon-secret-key\") pod \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.237989 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-logs\") pod \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.238025 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-scripts\") pod \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.238058 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5fc5\" (UniqueName: \"kubernetes.io/projected/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-kube-api-access-p5fc5\") pod \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.238090 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-config-data\") pod \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\" (UID: \"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15\") " Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.238432 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-logs" (OuterVolumeSpecName: "logs") pod "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" (UID: "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.243575 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" (UID: "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.244108 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-kube-api-access-p5fc5" (OuterVolumeSpecName: "kube-api-access-p5fc5") pod "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" (UID: "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15"). InnerVolumeSpecName "kube-api-access-p5fc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.264058 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-config-data" (OuterVolumeSpecName: "config-data") pod "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" (UID: "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.265154 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-scripts" (OuterVolumeSpecName: "scripts") pod "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" (UID: "3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.342272 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-logs\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.342301 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.342311 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5fc5\" (UniqueName: \"kubernetes.io/projected/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-kube-api-access-p5fc5\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.342514 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.342533 4741 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.497943 4741 generic.go:334] "Generic (PLEG): container finished" podID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerID="d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58" exitCode=137 Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.497980 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6468bdd6cc-7gxj7" event={"ID":"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15","Type":"ContainerDied","Data":"d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58"} Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.498007 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6468bdd6cc-7gxj7" event={"ID":"3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15","Type":"ContainerDied","Data":"147b2a17add310dc1ba9698a01d60668e61545de3ed3681619e4c89de726da3f"} Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.498023 4741 scope.go:117] "RemoveContainer" containerID="30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.498054 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6468bdd6cc-7gxj7" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.530991 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6468bdd6cc-7gxj7"] Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.540113 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6468bdd6cc-7gxj7"] Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.652009 4741 scope.go:117] "RemoveContainer" containerID="d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.677472 4741 scope.go:117] "RemoveContainer" containerID="30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4" Sep 29 20:50:31 crc kubenswrapper[4741]: E0929 20:50:31.677816 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4\": container with ID starting with 30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4 not found: ID does not exist" containerID="30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.677850 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4"} err="failed to get container status \"30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4\": rpc error: code = NotFound desc = could not find container \"30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4\": container with ID starting with 30b1699e2ecf27c71243c632cd9a42018980d8fc8955d092a83a81c335c918d4 not found: ID does not exist" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.677875 4741 scope.go:117] "RemoveContainer" containerID="d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58" Sep 29 20:50:31 crc kubenswrapper[4741]: E0929 20:50:31.678346 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58\": container with ID starting with d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58 not found: ID does not exist" containerID="d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58" Sep 29 20:50:31 crc kubenswrapper[4741]: I0929 20:50:31.678369 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58"} err="failed to get container status \"d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58\": rpc error: code = NotFound desc = could not find container \"d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58\": container with ID starting with d5bd3bee6c208c83d609adad43e65d7219c2d9fafce41f64da8b6d66a99e5e58 not found: ID does not exist" Sep 29 20:50:33 crc kubenswrapper[4741]: I0929 20:50:33.098172 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" path="/var/lib/kubelet/pods/3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15/volumes" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.211241 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5"] Sep 29 20:50:36 crc kubenswrapper[4741]: E0929 20:50:36.213563 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon-log" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.213597 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon-log" Sep 29 20:50:36 crc kubenswrapper[4741]: E0929 20:50:36.213643 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.213656 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.214022 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon-log" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.214056 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa2eebe-39ed-478d-84ca-f2fd4d1b8f15" containerName="horizon" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.216676 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.220907 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5"] Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.222588 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.248845 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g747q\" (UniqueName: \"kubernetes.io/projected/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-kube-api-access-g747q\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.248917 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.248949 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.351116 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g747q\" (UniqueName: \"kubernetes.io/projected/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-kube-api-access-g747q\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.351552 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.351771 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.352489 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.352581 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.376377 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g747q\" (UniqueName: \"kubernetes.io/projected/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-kube-api-access-g747q\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:36 crc kubenswrapper[4741]: I0929 20:50:36.544448 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:37 crc kubenswrapper[4741]: I0929 20:50:37.027299 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5"] Sep 29 20:50:37 crc kubenswrapper[4741]: I0929 20:50:37.570553 4741 generic.go:334] "Generic (PLEG): container finished" podID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerID="3ee5852f7bdef4345d098ce59008b6d6ed1d1a7262c1d9f405505fea85200a72" exitCode=0 Sep 29 20:50:37 crc kubenswrapper[4741]: I0929 20:50:37.570589 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" event={"ID":"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039","Type":"ContainerDied","Data":"3ee5852f7bdef4345d098ce59008b6d6ed1d1a7262c1d9f405505fea85200a72"} Sep 29 20:50:37 crc kubenswrapper[4741]: I0929 20:50:37.570610 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" event={"ID":"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039","Type":"ContainerStarted","Data":"97dd1cbac061a23fc0d4abe4a60694a9ae0c07e1ec758812fffc22dbbe189bd0"} Sep 29 20:50:38 crc kubenswrapper[4741]: I0929 20:50:38.085873 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:50:38 crc kubenswrapper[4741]: E0929 20:50:38.086756 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:50:39 crc kubenswrapper[4741]: I0929 20:50:39.585196 4741 generic.go:334] "Generic (PLEG): container finished" podID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerID="72b71cc45b28636889ec48c58145ed0f0b83e901aeaa6eb2e4ef340d64305abc" exitCode=0 Sep 29 20:50:39 crc kubenswrapper[4741]: I0929 20:50:39.585241 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" event={"ID":"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039","Type":"ContainerDied","Data":"72b71cc45b28636889ec48c58145ed0f0b83e901aeaa6eb2e4ef340d64305abc"} Sep 29 20:50:40 crc kubenswrapper[4741]: I0929 20:50:40.595259 4741 generic.go:334] "Generic (PLEG): container finished" podID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerID="f659e4b6ebe775d182238acfd97ff9eeaf74ce0d4e79694405ff2a18eb68f714" exitCode=0 Sep 29 20:50:40 crc kubenswrapper[4741]: I0929 20:50:40.595352 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" event={"ID":"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039","Type":"ContainerDied","Data":"f659e4b6ebe775d182238acfd97ff9eeaf74ce0d4e79694405ff2a18eb68f714"} Sep 29 20:50:41 crc kubenswrapper[4741]: I0929 20:50:41.965774 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.061228 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g747q\" (UniqueName: \"kubernetes.io/projected/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-kube-api-access-g747q\") pod \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.061603 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-util\") pod \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.061761 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-bundle\") pod \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\" (UID: \"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039\") " Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.063557 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-bundle" (OuterVolumeSpecName: "bundle") pod "ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" (UID: "ba55057a-3c2f-4ea0-b2c3-a9b1f896f039"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.069855 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-kube-api-access-g747q" (OuterVolumeSpecName: "kube-api-access-g747q") pod "ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" (UID: "ba55057a-3c2f-4ea0-b2c3-a9b1f896f039"). InnerVolumeSpecName "kube-api-access-g747q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.071706 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-util" (OuterVolumeSpecName: "util") pod "ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" (UID: "ba55057a-3c2f-4ea0-b2c3-a9b1f896f039"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.165068 4741 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-util\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.165334 4741 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.165347 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g747q\" (UniqueName: \"kubernetes.io/projected/ba55057a-3c2f-4ea0-b2c3-a9b1f896f039-kube-api-access-g747q\") on node \"crc\" DevicePath \"\"" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.613893 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" event={"ID":"ba55057a-3c2f-4ea0-b2c3-a9b1f896f039","Type":"ContainerDied","Data":"97dd1cbac061a23fc0d4abe4a60694a9ae0c07e1ec758812fffc22dbbe189bd0"} Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.613931 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97dd1cbac061a23fc0d4abe4a60694a9ae0c07e1ec758812fffc22dbbe189bd0" Sep 29 20:50:42 crc kubenswrapper[4741]: I0929 20:50:42.613949 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5" Sep 29 20:50:50 crc kubenswrapper[4741]: I0929 20:50:50.086352 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:50:50 crc kubenswrapper[4741]: E0929 20:50:50.087255 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.168395 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2"] Sep 29 20:50:53 crc kubenswrapper[4741]: E0929 20:50:53.169301 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerName="extract" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.169313 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerName="extract" Sep 29 20:50:53 crc kubenswrapper[4741]: E0929 20:50:53.169328 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerName="pull" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.169334 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerName="pull" Sep 29 20:50:53 crc kubenswrapper[4741]: E0929 20:50:53.169362 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerName="util" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.169368 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerName="util" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.170957 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba55057a-3c2f-4ea0-b2c3-a9b1f896f039" containerName="extract" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.171700 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.174039 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-9dcss" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.174083 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.174050 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.192938 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlwjm\" (UniqueName: \"kubernetes.io/projected/c6b5ed3e-0d65-49c3-b2af-5eec98429b23-kube-api-access-vlwjm\") pod \"obo-prometheus-operator-7c8cf85677-hzrg2\" (UID: \"c6b5ed3e-0d65-49c3-b2af-5eec98429b23\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.198603 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.294543 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlwjm\" (UniqueName: \"kubernetes.io/projected/c6b5ed3e-0d65-49c3-b2af-5eec98429b23-kube-api-access-vlwjm\") pod \"obo-prometheus-operator-7c8cf85677-hzrg2\" (UID: \"c6b5ed3e-0d65-49c3-b2af-5eec98429b23\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.303969 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.306171 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.354793 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-wj59d" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.355190 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.376514 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.390768 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.400758 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce0ae04d-8ff8-4668-8319-b9a955b7a6ec-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s\" (UID: \"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.400833 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce0ae04d-8ff8-4668-8319-b9a955b7a6ec-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s\" (UID: \"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.404485 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.404606 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.428754 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlwjm\" (UniqueName: \"kubernetes.io/projected/c6b5ed3e-0d65-49c3-b2af-5eec98429b23-kube-api-access-vlwjm\") pod \"obo-prometheus-operator-7c8cf85677-hzrg2\" (UID: \"c6b5ed3e-0d65-49c3-b2af-5eec98429b23\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.500939 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.502776 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce0ae04d-8ff8-4668-8319-b9a955b7a6ec-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s\" (UID: \"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.502835 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce0ae04d-8ff8-4668-8319-b9a955b7a6ec-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s\" (UID: \"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.502919 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51c125d4-27fa-4b7e-a99c-2253fb9e3e33-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-65492\" (UID: \"51c125d4-27fa-4b7e-a99c-2253fb9e3e33\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.502964 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51c125d4-27fa-4b7e-a99c-2253fb9e3e33-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-65492\" (UID: \"51c125d4-27fa-4b7e-a99c-2253fb9e3e33\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.507764 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce0ae04d-8ff8-4668-8319-b9a955b7a6ec-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s\" (UID: \"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.508811 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce0ae04d-8ff8-4668-8319-b9a955b7a6ec-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s\" (UID: \"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.543065 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-8kkkd"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.546904 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.550123 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-dhp4x" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.556463 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.563965 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-8kkkd"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.605189 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51c125d4-27fa-4b7e-a99c-2253fb9e3e33-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-65492\" (UID: \"51c125d4-27fa-4b7e-a99c-2253fb9e3e33\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.605252 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51c125d4-27fa-4b7e-a99c-2253fb9e3e33-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-65492\" (UID: \"51c125d4-27fa-4b7e-a99c-2253fb9e3e33\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.605328 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a791644a-1784-4f93-9772-85672174eafa-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-8kkkd\" (UID: \"a791644a-1784-4f93-9772-85672174eafa\") " pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.605412 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8mmq\" (UniqueName: \"kubernetes.io/projected/a791644a-1784-4f93-9772-85672174eafa-kube-api-access-l8mmq\") pod \"observability-operator-cc5f78dfc-8kkkd\" (UID: \"a791644a-1784-4f93-9772-85672174eafa\") " pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.608671 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51c125d4-27fa-4b7e-a99c-2253fb9e3e33-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-65492\" (UID: \"51c125d4-27fa-4b7e-a99c-2253fb9e3e33\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.610871 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51c125d4-27fa-4b7e-a99c-2253fb9e3e33-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6f89b4f898-65492\" (UID: \"51c125d4-27fa-4b7e-a99c-2253fb9e3e33\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.681754 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.712732 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8mmq\" (UniqueName: \"kubernetes.io/projected/a791644a-1784-4f93-9772-85672174eafa-kube-api-access-l8mmq\") pod \"observability-operator-cc5f78dfc-8kkkd\" (UID: \"a791644a-1784-4f93-9772-85672174eafa\") " pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.713030 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a791644a-1784-4f93-9772-85672174eafa-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-8kkkd\" (UID: \"a791644a-1784-4f93-9772-85672174eafa\") " pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.717631 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a791644a-1784-4f93-9772-85672174eafa-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-8kkkd\" (UID: \"a791644a-1784-4f93-9772-85672174eafa\") " pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.719787 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-m97dn"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.724855 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.727032 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-hv6tw" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.737008 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8mmq\" (UniqueName: \"kubernetes.io/projected/a791644a-1784-4f93-9772-85672174eafa-kube-api-access-l8mmq\") pod \"observability-operator-cc5f78dfc-8kkkd\" (UID: \"a791644a-1784-4f93-9772-85672174eafa\") " pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.758330 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-m97dn"] Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.772313 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.815755 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq565\" (UniqueName: \"kubernetes.io/projected/a12fd15d-08fb-4df4-842a-513513640697-kube-api-access-nq565\") pod \"perses-operator-54bc95c9fb-m97dn\" (UID: \"a12fd15d-08fb-4df4-842a-513513640697\") " pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.816130 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a12fd15d-08fb-4df4-842a-513513640697-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-m97dn\" (UID: \"a12fd15d-08fb-4df4-842a-513513640697\") " pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.917787 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a12fd15d-08fb-4df4-842a-513513640697-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-m97dn\" (UID: \"a12fd15d-08fb-4df4-842a-513513640697\") " pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.917937 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq565\" (UniqueName: \"kubernetes.io/projected/a12fd15d-08fb-4df4-842a-513513640697-kube-api-access-nq565\") pod \"perses-operator-54bc95c9fb-m97dn\" (UID: \"a12fd15d-08fb-4df4-842a-513513640697\") " pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.919170 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a12fd15d-08fb-4df4-842a-513513640697-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-m97dn\" (UID: \"a12fd15d-08fb-4df4-842a-513513640697\") " pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:53 crc kubenswrapper[4741]: I0929 20:50:53.945989 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq565\" (UniqueName: \"kubernetes.io/projected/a12fd15d-08fb-4df4-842a-513513640697-kube-api-access-nq565\") pod \"perses-operator-54bc95c9fb-m97dn\" (UID: \"a12fd15d-08fb-4df4-842a-513513640697\") " pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.011007 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.062173 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.288098 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2"] Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.390028 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s"] Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.547295 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492"] Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.729284 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-8kkkd"] Sep 29 20:50:54 crc kubenswrapper[4741]: W0929 20:50:54.733371 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda791644a_1784_4f93_9772_85672174eafa.slice/crio-fb4400dccdd070a263a0407f1715bca9113f2c23515f98c57f91c822571fb76c WatchSource:0}: Error finding container fb4400dccdd070a263a0407f1715bca9113f2c23515f98c57f91c822571fb76c: Status 404 returned error can't find the container with id fb4400dccdd070a263a0407f1715bca9113f2c23515f98c57f91c822571fb76c Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.795851 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" event={"ID":"c6b5ed3e-0d65-49c3-b2af-5eec98429b23","Type":"ContainerStarted","Data":"9ddb30f339fbf7db1f699a5b149a0ec164cbdab2d4ea3d74b12e3d8fcf8345b8"} Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.797634 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" event={"ID":"51c125d4-27fa-4b7e-a99c-2253fb9e3e33","Type":"ContainerStarted","Data":"7c33f03ac0e58fb229aab003164b7a14675a251dcb12838a1953ec36a6b7ccdd"} Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.799312 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" event={"ID":"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec","Type":"ContainerStarted","Data":"adb09faedf33c9c67d4c65adca9a5745c1f2ccaa8f41a72554e07537e4b4b856"} Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.800668 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" event={"ID":"a791644a-1784-4f93-9772-85672174eafa","Type":"ContainerStarted","Data":"fb4400dccdd070a263a0407f1715bca9113f2c23515f98c57f91c822571fb76c"} Sep 29 20:50:54 crc kubenswrapper[4741]: W0929 20:50:54.853326 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda12fd15d_08fb_4df4_842a_513513640697.slice/crio-5dcb40dd144355cd862de884b526c43e411bfec83f8065f9a7a0a2b3a6dcf9ce WatchSource:0}: Error finding container 5dcb40dd144355cd862de884b526c43e411bfec83f8065f9a7a0a2b3a6dcf9ce: Status 404 returned error can't find the container with id 5dcb40dd144355cd862de884b526c43e411bfec83f8065f9a7a0a2b3a6dcf9ce Sep 29 20:50:54 crc kubenswrapper[4741]: I0929 20:50:54.856001 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-m97dn"] Sep 29 20:50:55 crc kubenswrapper[4741]: I0929 20:50:55.811661 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" event={"ID":"a12fd15d-08fb-4df4-842a-513513640697","Type":"ContainerStarted","Data":"5dcb40dd144355cd862de884b526c43e411bfec83f8065f9a7a0a2b3a6dcf9ce"} Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.086125 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:51:03 crc kubenswrapper[4741]: E0929 20:51:03.086994 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.900003 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" event={"ID":"51c125d4-27fa-4b7e-a99c-2253fb9e3e33","Type":"ContainerStarted","Data":"c2f37b5d42d195a96935e261efe358b090f1460858d2cbedb0d8843af89d542c"} Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.902818 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" event={"ID":"ce0ae04d-8ff8-4668-8319-b9a955b7a6ec","Type":"ContainerStarted","Data":"54c577425ed42f33b73c8f73fc9cb0892c3dc630b7f4fca39c69397020f1cce3"} Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.904504 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" event={"ID":"a791644a-1784-4f93-9772-85672174eafa","Type":"ContainerStarted","Data":"5181f71c6640fa02104f338a724d42754196299392275f474964dd4ea97fe89f"} Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.905721 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.908043 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.908733 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" event={"ID":"c6b5ed3e-0d65-49c3-b2af-5eec98429b23","Type":"ContainerStarted","Data":"743f9a2bea3ef5c6ebee1b07b34493f1a40de974ced538181c5a92ff19c62782"} Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.911450 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" event={"ID":"a12fd15d-08fb-4df4-842a-513513640697","Type":"ContainerStarted","Data":"c441066d843867cfad2a635abb8468cdb6a195e587076ad1e620aa7b7fa62d1f"} Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.911639 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.947345 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-65492" podStartSLOduration=2.912751282 podStartE2EDuration="10.947318788s" podCreationTimestamp="2025-09-29 20:50:53 +0000 UTC" firstStartedPulling="2025-09-29 20:50:54.537897306 +0000 UTC m=+6096.185686638" lastFinishedPulling="2025-09-29 20:51:02.572464812 +0000 UTC m=+6104.220254144" observedRunningTime="2025-09-29 20:51:03.928735822 +0000 UTC m=+6105.576525164" watchObservedRunningTime="2025-09-29 20:51:03.947318788 +0000 UTC m=+6105.595108130" Sep 29 20:51:03 crc kubenswrapper[4741]: I0929 20:51:03.973569 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s" podStartSLOduration=2.854569218 podStartE2EDuration="10.973549823s" podCreationTimestamp="2025-09-29 20:50:53 +0000 UTC" firstStartedPulling="2025-09-29 20:50:54.453774377 +0000 UTC m=+6096.101563719" lastFinishedPulling="2025-09-29 20:51:02.572754992 +0000 UTC m=+6104.220544324" observedRunningTime="2025-09-29 20:51:03.957638092 +0000 UTC m=+6105.605427424" watchObservedRunningTime="2025-09-29 20:51:03.973549823 +0000 UTC m=+6105.621339155" Sep 29 20:51:04 crc kubenswrapper[4741]: I0929 20:51:04.010774 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hzrg2" podStartSLOduration=2.754672703 podStartE2EDuration="11.010756635s" podCreationTimestamp="2025-09-29 20:50:53 +0000 UTC" firstStartedPulling="2025-09-29 20:50:54.318126185 +0000 UTC m=+6095.965915517" lastFinishedPulling="2025-09-29 20:51:02.574210087 +0000 UTC m=+6104.221999449" observedRunningTime="2025-09-29 20:51:04.004010843 +0000 UTC m=+6105.651800175" watchObservedRunningTime="2025-09-29 20:51:04.010756635 +0000 UTC m=+6105.658545967" Sep 29 20:51:04 crc kubenswrapper[4741]: I0929 20:51:04.035759 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" podStartSLOduration=3.286721088 podStartE2EDuration="11.035736472s" podCreationTimestamp="2025-09-29 20:50:53 +0000 UTC" firstStartedPulling="2025-09-29 20:50:54.855704994 +0000 UTC m=+6096.503494316" lastFinishedPulling="2025-09-29 20:51:02.604720368 +0000 UTC m=+6104.252509700" observedRunningTime="2025-09-29 20:51:04.026116409 +0000 UTC m=+6105.673905741" watchObservedRunningTime="2025-09-29 20:51:04.035736472 +0000 UTC m=+6105.683525804" Sep 29 20:51:09 crc kubenswrapper[4741]: I0929 20:51:09.043037 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-8kkkd" podStartSLOduration=8.155052205 podStartE2EDuration="16.043019486s" podCreationTimestamp="2025-09-29 20:50:53 +0000 UTC" firstStartedPulling="2025-09-29 20:50:54.736006244 +0000 UTC m=+6096.383795576" lastFinishedPulling="2025-09-29 20:51:02.623973525 +0000 UTC m=+6104.271762857" observedRunningTime="2025-09-29 20:51:04.07345938 +0000 UTC m=+6105.721248712" watchObservedRunningTime="2025-09-29 20:51:09.043019486 +0000 UTC m=+6110.690808818" Sep 29 20:51:09 crc kubenswrapper[4741]: I0929 20:51:09.047720 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-bf56t"] Sep 29 20:51:09 crc kubenswrapper[4741]: I0929 20:51:09.055246 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-98rj6"] Sep 29 20:51:09 crc kubenswrapper[4741]: I0929 20:51:09.062973 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-98rj6"] Sep 29 20:51:09 crc kubenswrapper[4741]: I0929 20:51:09.070192 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-bf56t"] Sep 29 20:51:09 crc kubenswrapper[4741]: I0929 20:51:09.096357 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6" path="/var/lib/kubelet/pods/1f2c8864-ff5d-48f8-bfcd-1f1104ffe9a6/volumes" Sep 29 20:51:09 crc kubenswrapper[4741]: I0929 20:51:09.098286 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3dce1e-ade4-4090-a9c5-97238453667f" path="/var/lib/kubelet/pods/9a3dce1e-ade4-4090-a9c5-97238453667f/volumes" Sep 29 20:51:10 crc kubenswrapper[4741]: I0929 20:51:10.029601 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-djpqv"] Sep 29 20:51:10 crc kubenswrapper[4741]: I0929 20:51:10.048826 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-djpqv"] Sep 29 20:51:11 crc kubenswrapper[4741]: I0929 20:51:11.097336 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb77939-b991-416f-be87-085e7d762af6" path="/var/lib/kubelet/pods/abb77939-b991-416f-be87-085e7d762af6/volumes" Sep 29 20:51:14 crc kubenswrapper[4741]: I0929 20:51:14.064326 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-m97dn" Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.085901 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:51:16 crc kubenswrapper[4741]: E0929 20:51:16.086436 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.854510 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.854764 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" containerName="openstackclient" containerID="cri-o://eaef049f3ddd3bc029b384e13683dd571330298cf39a43da1c0596d11b82e9bb" gracePeriod=2 Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.863158 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.896424 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 29 20:51:16 crc kubenswrapper[4741]: E0929 20:51:16.896804 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" containerName="openstackclient" Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.896823 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" containerName="openstackclient" Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.897014 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" containerName="openstackclient" Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.897665 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.912464 4741 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" podUID="786fbbc4-3a41-4732-8ec1-c948fa4a346f" Sep 29 20:51:16 crc kubenswrapper[4741]: I0929 20:51:16.920354 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.010782 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/786fbbc4-3a41-4732-8ec1-c948fa4a346f-openstack-config-secret\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.010970 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvsk4\" (UniqueName: \"kubernetes.io/projected/786fbbc4-3a41-4732-8ec1-c948fa4a346f-kube-api-access-bvsk4\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.011194 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/786fbbc4-3a41-4732-8ec1-c948fa4a346f-openstack-config\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.098169 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.099415 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.110602 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.113752 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvsk4\" (UniqueName: \"kubernetes.io/projected/786fbbc4-3a41-4732-8ec1-c948fa4a346f-kube-api-access-bvsk4\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.113823 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/786fbbc4-3a41-4732-8ec1-c948fa4a346f-openstack-config\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.113890 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/786fbbc4-3a41-4732-8ec1-c948fa4a346f-openstack-config-secret\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.118083 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-m5cl6" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.142145 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/786fbbc4-3a41-4732-8ec1-c948fa4a346f-openstack-config-secret\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.147044 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/786fbbc4-3a41-4732-8ec1-c948fa4a346f-openstack-config\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.161311 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvsk4\" (UniqueName: \"kubernetes.io/projected/786fbbc4-3a41-4732-8ec1-c948fa4a346f-kube-api-access-bvsk4\") pod \"openstackclient\" (UID: \"786fbbc4-3a41-4732-8ec1-c948fa4a346f\") " pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.213814 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.218834 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7p8n\" (UniqueName: \"kubernetes.io/projected/c998616e-1aed-4b70-a3fe-05d625649fa4-kube-api-access-v7p8n\") pod \"kube-state-metrics-0\" (UID: \"c998616e-1aed-4b70-a3fe-05d625649fa4\") " pod="openstack/kube-state-metrics-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.325623 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7p8n\" (UniqueName: \"kubernetes.io/projected/c998616e-1aed-4b70-a3fe-05d625649fa4-kube-api-access-v7p8n\") pod \"kube-state-metrics-0\" (UID: \"c998616e-1aed-4b70-a3fe-05d625649fa4\") " pod="openstack/kube-state-metrics-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.366250 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7p8n\" (UniqueName: \"kubernetes.io/projected/c998616e-1aed-4b70-a3fe-05d625649fa4-kube-api-access-v7p8n\") pod \"kube-state-metrics-0\" (UID: \"c998616e-1aed-4b70-a3fe-05d625649fa4\") " pod="openstack/kube-state-metrics-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.414300 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.764058 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.766603 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.771033 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.771310 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.772137 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-knqz8" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.772305 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.789110 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.848496 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5178344b-4d12-4044-bf16-ea6eb56dc4f7-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.848802 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5178344b-4d12-4044-bf16-ea6eb56dc4f7-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.848831 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t79x\" (UniqueName: \"kubernetes.io/projected/5178344b-4d12-4044-bf16-ea6eb56dc4f7-kube-api-access-5t79x\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.849068 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5178344b-4d12-4044-bf16-ea6eb56dc4f7-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.849155 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5178344b-4d12-4044-bf16-ea6eb56dc4f7-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.849308 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5178344b-4d12-4044-bf16-ea6eb56dc4f7-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.951136 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5178344b-4d12-4044-bf16-ea6eb56dc4f7-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.951209 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5178344b-4d12-4044-bf16-ea6eb56dc4f7-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.951235 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t79x\" (UniqueName: \"kubernetes.io/projected/5178344b-4d12-4044-bf16-ea6eb56dc4f7-kube-api-access-5t79x\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.951288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5178344b-4d12-4044-bf16-ea6eb56dc4f7-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.951316 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5178344b-4d12-4044-bf16-ea6eb56dc4f7-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.951361 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5178344b-4d12-4044-bf16-ea6eb56dc4f7-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.953081 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5178344b-4d12-4044-bf16-ea6eb56dc4f7-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.959657 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5178344b-4d12-4044-bf16-ea6eb56dc4f7-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.961124 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5178344b-4d12-4044-bf16-ea6eb56dc4f7-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.975080 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5178344b-4d12-4044-bf16-ea6eb56dc4f7-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.975904 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5178344b-4d12-4044-bf16-ea6eb56dc4f7-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:17 crc kubenswrapper[4741]: I0929 20:51:17.991229 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t79x\" (UniqueName: \"kubernetes.io/projected/5178344b-4d12-4044-bf16-ea6eb56dc4f7-kube-api-access-5t79x\") pod \"alertmanager-metric-storage-0\" (UID: \"5178344b-4d12-4044-bf16-ea6eb56dc4f7\") " pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.031212 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.086013 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.098914 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.426443 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.440840 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.446993 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.447238 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.463930 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.464224 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.464600 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-sm8d7" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.490914 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.495761 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.586705 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.587055 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.587111 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.587147 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.587217 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.587247 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-config\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.587261 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.587282 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nllq2\" (UniqueName: \"kubernetes.io/projected/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-kube-api-access-nllq2\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693049 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693178 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693211 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-config\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693245 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693272 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nllq2\" (UniqueName: \"kubernetes.io/projected/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-kube-api-access-nllq2\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693322 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693380 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.693456 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.697204 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.706867 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.714133 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.714433 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-config\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.715024 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.726896 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nllq2\" (UniqueName: \"kubernetes.io/projected/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-kube-api-access-nllq2\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.729009 4741 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.729033 4741 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5eca29a852bb14942aabaf6d48d7f473fc34e32edb9a51f1a96fb138cf64fbc5/globalmount\"" pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.729091 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8396f12c-5cd2-4db4-b2cb-0fa25e63efaf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:18 crc kubenswrapper[4741]: I0929 20:51:18.865852 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.010068 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-244fae9d-c718-46cc-bfd9-6d622558aa6c\") pod \"prometheus-metric-storage-0\" (UID: \"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf\") " pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.188327 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-sm8d7" Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.191242 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.205801 4741 generic.go:334] "Generic (PLEG): container finished" podID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" containerID="eaef049f3ddd3bc029b384e13683dd571330298cf39a43da1c0596d11b82e9bb" exitCode=137 Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.221608 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5178344b-4d12-4044-bf16-ea6eb56dc4f7","Type":"ContainerStarted","Data":"8fcb5c278176361645c0b56145da13ca0db06caaf4ef43ceeb7fdb1b97e025f2"} Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.226966 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"786fbbc4-3a41-4732-8ec1-c948fa4a346f","Type":"ContainerStarted","Data":"638d5898b5f0aa97c92a809f16b2b725dbef01b5c24bccb5b84f94280b8e8cd7"} Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.249882 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c998616e-1aed-4b70-a3fe-05d625649fa4","Type":"ContainerStarted","Data":"4c24f215a708acaac9ee2ec7dcbfea998fae7d2642d8935629c5b5b67c9f5619"} Sep 29 20:51:19 crc kubenswrapper[4741]: I0929 20:51:19.810082 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 29 20:51:19 crc kubenswrapper[4741]: W0929 20:51:19.846750 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8396f12c_5cd2_4db4_b2cb_0fa25e63efaf.slice/crio-cf763d126a878b8380c2285a4d326ab558454ad0c7950d675ad0a9b6a45a2124 WatchSource:0}: Error finding container cf763d126a878b8380c2285a4d326ab558454ad0c7950d675ad0a9b6a45a2124: Status 404 returned error can't find the container with id cf763d126a878b8380c2285a4d326ab558454ad0c7950d675ad0a9b6a45a2124 Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.036460 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-dd71-account-create-9spsc"] Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.044191 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-6963-account-create-f2nw9"] Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.075436 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-100e-account-create-fd4xg"] Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.099924 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-dd71-account-create-9spsc"] Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.117399 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-100e-account-create-fd4xg"] Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.124134 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-6963-account-create-f2nw9"] Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.272748 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"786fbbc4-3a41-4732-8ec1-c948fa4a346f","Type":"ContainerStarted","Data":"7b38f4667b23e5d00abe29458275ba312bb520320f343e52f8bb130bfbc9bab7"} Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.279131 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf","Type":"ContainerStarted","Data":"cf763d126a878b8380c2285a4d326ab558454ad0c7950d675ad0a9b6a45a2124"} Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.453629 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.456350 4741 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" podUID="786fbbc4-3a41-4732-8ec1-c948fa4a346f" Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.461855 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgtmp\" (UniqueName: \"kubernetes.io/projected/65ecdd59-688a-45c4-bd78-a4a0a27f338c-kube-api-access-qgtmp\") pod \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.461982 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config-secret\") pod \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.462007 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config\") pod \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\" (UID: \"65ecdd59-688a-45c4-bd78-a4a0a27f338c\") " Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.468700 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ecdd59-688a-45c4-bd78-a4a0a27f338c-kube-api-access-qgtmp" (OuterVolumeSpecName: "kube-api-access-qgtmp") pod "65ecdd59-688a-45c4-bd78-a4a0a27f338c" (UID: "65ecdd59-688a-45c4-bd78-a4a0a27f338c"). InnerVolumeSpecName "kube-api-access-qgtmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.490705 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "65ecdd59-688a-45c4-bd78-a4a0a27f338c" (UID: "65ecdd59-688a-45c4-bd78-a4a0a27f338c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.540420 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "65ecdd59-688a-45c4-bd78-a4a0a27f338c" (UID: "65ecdd59-688a-45c4-bd78-a4a0a27f338c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.562994 4741 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.563026 4741 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/65ecdd59-688a-45c4-bd78-a4a0a27f338c-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:51:20 crc kubenswrapper[4741]: I0929 20:51:20.563037 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgtmp\" (UniqueName: \"kubernetes.io/projected/65ecdd59-688a-45c4-bd78-a4a0a27f338c-kube-api-access-qgtmp\") on node \"crc\" DevicePath \"\"" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.158639 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="088c9042-fb78-4b36-a9ff-7c5ae4747861" path="/var/lib/kubelet/pods/088c9042-fb78-4b36-a9ff-7c5ae4747861/volumes" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.162605 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57de32ae-a8b9-4ab7-b8e5-6cc475543766" path="/var/lib/kubelet/pods/57de32ae-a8b9-4ab7-b8e5-6cc475543766/volumes" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.174854 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="603b9f56-b1fc-4fdc-bf43-3cc338c737e2" path="/var/lib/kubelet/pods/603b9f56-b1fc-4fdc-bf43-3cc338c737e2/volumes" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.176446 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" path="/var/lib/kubelet/pods/65ecdd59-688a-45c4-bd78-a4a0a27f338c/volumes" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.295859 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c998616e-1aed-4b70-a3fe-05d625649fa4","Type":"ContainerStarted","Data":"e8ca5d4827906faa5747134bad31da61f3f52e5e16d14f5f0df23f23c8d0baa0"} Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.297242 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.298603 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.298628 4741 scope.go:117] "RemoveContainer" containerID="eaef049f3ddd3bc029b384e13683dd571330298cf39a43da1c0596d11b82e9bb" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.320204 4741 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="65ecdd59-688a-45c4-bd78-a4a0a27f338c" podUID="786fbbc4-3a41-4732-8ec1-c948fa4a346f" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.322282 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=5.32227117 podStartE2EDuration="5.32227117s" podCreationTimestamp="2025-09-29 20:51:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:51:20.295849047 +0000 UTC m=+6121.943638389" watchObservedRunningTime="2025-09-29 20:51:21.32227117 +0000 UTC m=+6122.970060502" Sep 29 20:51:21 crc kubenswrapper[4741]: I0929 20:51:21.328168 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.07211316 podStartE2EDuration="4.328153665s" podCreationTimestamp="2025-09-29 20:51:17 +0000 UTC" firstStartedPulling="2025-09-29 20:51:18.105563323 +0000 UTC m=+6119.753352655" lastFinishedPulling="2025-09-29 20:51:20.361603838 +0000 UTC m=+6122.009393160" observedRunningTime="2025-09-29 20:51:21.314421452 +0000 UTC m=+6122.962210794" watchObservedRunningTime="2025-09-29 20:51:21.328153665 +0000 UTC m=+6122.975942997" Sep 29 20:51:26 crc kubenswrapper[4741]: I0929 20:51:26.365732 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf","Type":"ContainerStarted","Data":"8f01a1d919676d24d77305ec6c54988df8e204747db59c3df1818d81d9183f21"} Sep 29 20:51:26 crc kubenswrapper[4741]: I0929 20:51:26.369238 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5178344b-4d12-4044-bf16-ea6eb56dc4f7","Type":"ContainerStarted","Data":"bb444109b2d7f08a11e8c241a8404342b8a28c05efa5cc23b05dcdfb610a65fa"} Sep 29 20:51:27 crc kubenswrapper[4741]: I0929 20:51:27.418583 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 29 20:51:28 crc kubenswrapper[4741]: I0929 20:51:28.945154 4741 scope.go:117] "RemoveContainer" containerID="2cff54f16bbf7a47c92147596e8cb926ff57cb74f9ca667f207808920f49a0ec" Sep 29 20:51:28 crc kubenswrapper[4741]: I0929 20:51:28.971811 4741 scope.go:117] "RemoveContainer" containerID="aa0394e84fc4e7abd1d6df3213105e0e1d9f30a587b9fee62406ef5c2ea5761a" Sep 29 20:51:29 crc kubenswrapper[4741]: I0929 20:51:29.059700 4741 scope.go:117] "RemoveContainer" containerID="4a3567c4a021b49b7eb2de83f7a86b6856c33facca21427f0f7eae32b71679a2" Sep 29 20:51:29 crc kubenswrapper[4741]: I0929 20:51:29.096186 4741 scope.go:117] "RemoveContainer" containerID="6938ce916b4bebf042498dcf9de1535f9e718aed6efaa9d6936cdae21365486c" Sep 29 20:51:29 crc kubenswrapper[4741]: I0929 20:51:29.136569 4741 scope.go:117] "RemoveContainer" containerID="4a662aa200f12dc8733ec49f7e138748b3f8a8dcb88cf0cdbfafa7d7d06b66b8" Sep 29 20:51:29 crc kubenswrapper[4741]: I0929 20:51:29.185937 4741 scope.go:117] "RemoveContainer" containerID="b332dc715b490f8a47542561f3fc98534e0805d24bd11a3cdfed6228dbf21eb6" Sep 29 20:51:31 crc kubenswrapper[4741]: I0929 20:51:31.087105 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:51:31 crc kubenswrapper[4741]: E0929 20:51:31.088277 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:51:33 crc kubenswrapper[4741]: I0929 20:51:33.462126 4741 generic.go:334] "Generic (PLEG): container finished" podID="8396f12c-5cd2-4db4-b2cb-0fa25e63efaf" containerID="8f01a1d919676d24d77305ec6c54988df8e204747db59c3df1818d81d9183f21" exitCode=0 Sep 29 20:51:33 crc kubenswrapper[4741]: I0929 20:51:33.462237 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf","Type":"ContainerDied","Data":"8f01a1d919676d24d77305ec6c54988df8e204747db59c3df1818d81d9183f21"} Sep 29 20:51:33 crc kubenswrapper[4741]: I0929 20:51:33.464068 4741 generic.go:334] "Generic (PLEG): container finished" podID="5178344b-4d12-4044-bf16-ea6eb56dc4f7" containerID="bb444109b2d7f08a11e8c241a8404342b8a28c05efa5cc23b05dcdfb610a65fa" exitCode=0 Sep 29 20:51:33 crc kubenswrapper[4741]: I0929 20:51:33.464116 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5178344b-4d12-4044-bf16-ea6eb56dc4f7","Type":"ContainerDied","Data":"bb444109b2d7f08a11e8c241a8404342b8a28c05efa5cc23b05dcdfb610a65fa"} Sep 29 20:51:35 crc kubenswrapper[4741]: I0929 20:51:35.034887 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzs57"] Sep 29 20:51:35 crc kubenswrapper[4741]: I0929 20:51:35.045485 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzs57"] Sep 29 20:51:35 crc kubenswrapper[4741]: I0929 20:51:35.097413 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089efa06-d16d-4604-8d3e-6f7b2c427dd8" path="/var/lib/kubelet/pods/089efa06-d16d-4604-8d3e-6f7b2c427dd8/volumes" Sep 29 20:51:37 crc kubenswrapper[4741]: I0929 20:51:37.504082 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5178344b-4d12-4044-bf16-ea6eb56dc4f7","Type":"ContainerStarted","Data":"1aec573a7630c895124685247f54cfa45b8c71d1f0846443cee7d512f714f2ed"} Sep 29 20:51:41 crc kubenswrapper[4741]: I0929 20:51:41.555428 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf","Type":"ContainerStarted","Data":"cd11880dc479c3c5f78e00b01a33cefdd2de15f4900d6265e0097b78b5051381"} Sep 29 20:51:41 crc kubenswrapper[4741]: I0929 20:51:41.558354 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5178344b-4d12-4044-bf16-ea6eb56dc4f7","Type":"ContainerStarted","Data":"2df1cd140f7eccf5e113f487bd3bc7f260f38ee82f1cf22241fa729c784bf2fb"} Sep 29 20:51:41 crc kubenswrapper[4741]: I0929 20:51:41.558786 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:41 crc kubenswrapper[4741]: I0929 20:51:41.561508 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Sep 29 20:51:41 crc kubenswrapper[4741]: I0929 20:51:41.609755 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=7.248793289 podStartE2EDuration="24.609737661s" podCreationTimestamp="2025-09-29 20:51:17 +0000 UTC" firstStartedPulling="2025-09-29 20:51:18.966603108 +0000 UTC m=+6120.614392440" lastFinishedPulling="2025-09-29 20:51:36.32754745 +0000 UTC m=+6137.975336812" observedRunningTime="2025-09-29 20:51:41.586739046 +0000 UTC m=+6143.234528378" watchObservedRunningTime="2025-09-29 20:51:41.609737661 +0000 UTC m=+6143.257526993" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.596257 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf","Type":"ContainerStarted","Data":"145773fce7b6ca46e141e18c695eca57dfd275021b6c02aa120c73d84c3cdcbb"} Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.676975 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qkkxf"] Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.679484 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.691157 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qkkxf"] Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.722352 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-utilities\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.722532 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ntzp\" (UniqueName: \"kubernetes.io/projected/3e1c7427-d041-4489-89a1-e5b44ac89b30-kube-api-access-6ntzp\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.722596 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-catalog-content\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.824262 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-utilities\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.824355 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ntzp\" (UniqueName: \"kubernetes.io/projected/3e1c7427-d041-4489-89a1-e5b44ac89b30-kube-api-access-6ntzp\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.824406 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-catalog-content\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.824880 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-catalog-content\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.824988 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-utilities\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:44 crc kubenswrapper[4741]: I0929 20:51:44.852283 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ntzp\" (UniqueName: \"kubernetes.io/projected/3e1c7427-d041-4489-89a1-e5b44ac89b30-kube-api-access-6ntzp\") pod \"community-operators-qkkxf\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:45 crc kubenswrapper[4741]: I0929 20:51:44.999952 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:45 crc kubenswrapper[4741]: I0929 20:51:45.086028 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:51:45 crc kubenswrapper[4741]: E0929 20:51:45.086334 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:51:45 crc kubenswrapper[4741]: I0929 20:51:45.665820 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qkkxf"] Sep 29 20:51:45 crc kubenswrapper[4741]: W0929 20:51:45.685948 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e1c7427_d041_4489_89a1_e5b44ac89b30.slice/crio-a1b18cb197f4ade37664d97735e31b1c086e08847f62888197b29d9f578b2c03 WatchSource:0}: Error finding container a1b18cb197f4ade37664d97735e31b1c086e08847f62888197b29d9f578b2c03: Status 404 returned error can't find the container with id a1b18cb197f4ade37664d97735e31b1c086e08847f62888197b29d9f578b2c03 Sep 29 20:51:46 crc kubenswrapper[4741]: I0929 20:51:46.613673 4741 generic.go:334] "Generic (PLEG): container finished" podID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerID="6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6" exitCode=0 Sep 29 20:51:46 crc kubenswrapper[4741]: I0929 20:51:46.614997 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkkxf" event={"ID":"3e1c7427-d041-4489-89a1-e5b44ac89b30","Type":"ContainerDied","Data":"6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6"} Sep 29 20:51:46 crc kubenswrapper[4741]: I0929 20:51:46.615028 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkkxf" event={"ID":"3e1c7427-d041-4489-89a1-e5b44ac89b30","Type":"ContainerStarted","Data":"a1b18cb197f4ade37664d97735e31b1c086e08847f62888197b29d9f578b2c03"} Sep 29 20:51:48 crc kubenswrapper[4741]: I0929 20:51:48.638186 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkkxf" event={"ID":"3e1c7427-d041-4489-89a1-e5b44ac89b30","Type":"ContainerStarted","Data":"b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245"} Sep 29 20:51:48 crc kubenswrapper[4741]: I0929 20:51:48.641268 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8396f12c-5cd2-4db4-b2cb-0fa25e63efaf","Type":"ContainerStarted","Data":"a6816fdb786e33f57d2967250185e472e6ce0b3da0e6089b9daba2f1ff783dec"} Sep 29 20:51:48 crc kubenswrapper[4741]: I0929 20:51:48.685039 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=4.103183941 podStartE2EDuration="31.685019468s" podCreationTimestamp="2025-09-29 20:51:17 +0000 UTC" firstStartedPulling="2025-09-29 20:51:19.853962171 +0000 UTC m=+6121.501751503" lastFinishedPulling="2025-09-29 20:51:47.435797698 +0000 UTC m=+6149.083587030" observedRunningTime="2025-09-29 20:51:48.683578492 +0000 UTC m=+6150.331367864" watchObservedRunningTime="2025-09-29 20:51:48.685019468 +0000 UTC m=+6150.332808800" Sep 29 20:51:49 crc kubenswrapper[4741]: I0929 20:51:49.192416 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:49 crc kubenswrapper[4741]: I0929 20:51:49.192495 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:49 crc kubenswrapper[4741]: I0929 20:51:49.196825 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:49 crc kubenswrapper[4741]: I0929 20:51:49.661182 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Sep 29 20:51:50 crc kubenswrapper[4741]: I0929 20:51:50.668480 4741 generic.go:334] "Generic (PLEG): container finished" podID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerID="b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245" exitCode=0 Sep 29 20:51:50 crc kubenswrapper[4741]: I0929 20:51:50.668532 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkkxf" event={"ID":"3e1c7427-d041-4489-89a1-e5b44ac89b30","Type":"ContainerDied","Data":"b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245"} Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.399974 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.406616 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.410471 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.410886 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.448565 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.481493 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kdbb\" (UniqueName: \"kubernetes.io/projected/27f1d9d0-0481-4f24-840b-4040cbf232e5-kube-api-access-2kdbb\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.481547 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-scripts\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.481603 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.481878 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-log-httpd\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.482003 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-run-httpd\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.482082 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-config-data\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.482202 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583201 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583304 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-log-httpd\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583346 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-run-httpd\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583375 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-config-data\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583441 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583514 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kdbb\" (UniqueName: \"kubernetes.io/projected/27f1d9d0-0481-4f24-840b-4040cbf232e5-kube-api-access-2kdbb\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583554 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-scripts\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.583672 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-log-httpd\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.584666 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-run-httpd\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.589023 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.603836 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-config-data\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.608812 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-scripts\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.609162 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.610718 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kdbb\" (UniqueName: \"kubernetes.io/projected/27f1d9d0-0481-4f24-840b-4040cbf232e5-kube-api-access-2kdbb\") pod \"ceilometer-0\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " pod="openstack/ceilometer-0" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.686234 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkkxf" event={"ID":"3e1c7427-d041-4489-89a1-e5b44ac89b30","Type":"ContainerStarted","Data":"1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc"} Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.715160 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qkkxf" podStartSLOduration=3.812120379 podStartE2EDuration="7.715139118s" podCreationTimestamp="2025-09-29 20:51:44 +0000 UTC" firstStartedPulling="2025-09-29 20:51:47.363635346 +0000 UTC m=+6149.011424678" lastFinishedPulling="2025-09-29 20:51:51.266654095 +0000 UTC m=+6152.914443417" observedRunningTime="2025-09-29 20:51:51.704488293 +0000 UTC m=+6153.352277625" watchObservedRunningTime="2025-09-29 20:51:51.715139118 +0000 UTC m=+6153.362928450" Sep 29 20:51:51 crc kubenswrapper[4741]: I0929 20:51:51.738455 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:51:52 crc kubenswrapper[4741]: I0929 20:51:52.306104 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:51:52 crc kubenswrapper[4741]: W0929 20:51:52.308842 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27f1d9d0_0481_4f24_840b_4040cbf232e5.slice/crio-a45d5ce8f638953112136c50f1c32881ba8eb9da434c7cd327f31d814809355e WatchSource:0}: Error finding container a45d5ce8f638953112136c50f1c32881ba8eb9da434c7cd327f31d814809355e: Status 404 returned error can't find the container with id a45d5ce8f638953112136c50f1c32881ba8eb9da434c7cd327f31d814809355e Sep 29 20:51:52 crc kubenswrapper[4741]: I0929 20:51:52.696930 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerStarted","Data":"a45d5ce8f638953112136c50f1c32881ba8eb9da434c7cd327f31d814809355e"} Sep 29 20:51:53 crc kubenswrapper[4741]: I0929 20:51:53.706709 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerStarted","Data":"1862b05a3f863ba3f54ed86d164e1cc1f8ea4f460f984abf5ef8f72ce7b83f8b"} Sep 29 20:51:54 crc kubenswrapper[4741]: I0929 20:51:54.070820 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-84tzh"] Sep 29 20:51:54 crc kubenswrapper[4741]: I0929 20:51:54.082157 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-84tzh"] Sep 29 20:51:54 crc kubenswrapper[4741]: I0929 20:51:54.717022 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerStarted","Data":"22dc64af77da9960a0f5093954cd0bbecc8a233de0e6c9b2763c029c7517285f"} Sep 29 20:51:54 crc kubenswrapper[4741]: I0929 20:51:54.717360 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerStarted","Data":"c48f8b84d098762b1900b081283294ca4eaf14daf992db45cf8dfc9e10eebd4c"} Sep 29 20:51:55 crc kubenswrapper[4741]: I0929 20:51:55.001057 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:55 crc kubenswrapper[4741]: I0929 20:51:55.002219 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:55 crc kubenswrapper[4741]: I0929 20:51:55.060642 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:55 crc kubenswrapper[4741]: I0929 20:51:55.101693 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e" path="/var/lib/kubelet/pods/bebfdfb5-fc62-4dfc-ac1e-d7f1815bc29e/volumes" Sep 29 20:51:56 crc kubenswrapper[4741]: I0929 20:51:56.026177 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd5mn"] Sep 29 20:51:56 crc kubenswrapper[4741]: I0929 20:51:56.034884 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd5mn"] Sep 29 20:51:56 crc kubenswrapper[4741]: I0929 20:51:56.783514 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:56 crc kubenswrapper[4741]: I0929 20:51:56.837331 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qkkxf"] Sep 29 20:51:57 crc kubenswrapper[4741]: I0929 20:51:57.099366 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89fed09c-7c9c-40c8-82e3-dca8f90f0ff2" path="/var/lib/kubelet/pods/89fed09c-7c9c-40c8-82e3-dca8f90f0ff2/volumes" Sep 29 20:51:57 crc kubenswrapper[4741]: I0929 20:51:57.754356 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerStarted","Data":"0dbd4f9284d78e9080cf849ed6601ea67af721c050d29358648f8838411d35ae"} Sep 29 20:51:57 crc kubenswrapper[4741]: I0929 20:51:57.789349 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.623220265 podStartE2EDuration="6.7893171s" podCreationTimestamp="2025-09-29 20:51:51 +0000 UTC" firstStartedPulling="2025-09-29 20:51:52.311166347 +0000 UTC m=+6153.958955679" lastFinishedPulling="2025-09-29 20:51:56.477263182 +0000 UTC m=+6158.125052514" observedRunningTime="2025-09-29 20:51:57.782119833 +0000 UTC m=+6159.429909165" watchObservedRunningTime="2025-09-29 20:51:57.7893171 +0000 UTC m=+6159.437106432" Sep 29 20:51:58 crc kubenswrapper[4741]: I0929 20:51:58.767298 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qkkxf" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="registry-server" containerID="cri-o://1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc" gracePeriod=2 Sep 29 20:51:58 crc kubenswrapper[4741]: I0929 20:51:58.767336 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.092889 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:51:59 crc kubenswrapper[4741]: E0929 20:51:59.095123 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.345752 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.367073 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-catalog-content\") pod \"3e1c7427-d041-4489-89a1-e5b44ac89b30\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.367129 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ntzp\" (UniqueName: \"kubernetes.io/projected/3e1c7427-d041-4489-89a1-e5b44ac89b30-kube-api-access-6ntzp\") pod \"3e1c7427-d041-4489-89a1-e5b44ac89b30\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.367198 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-utilities\") pod \"3e1c7427-d041-4489-89a1-e5b44ac89b30\" (UID: \"3e1c7427-d041-4489-89a1-e5b44ac89b30\") " Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.368505 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-utilities" (OuterVolumeSpecName: "utilities") pod "3e1c7427-d041-4489-89a1-e5b44ac89b30" (UID: "3e1c7427-d041-4489-89a1-e5b44ac89b30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.375712 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e1c7427-d041-4489-89a1-e5b44ac89b30-kube-api-access-6ntzp" (OuterVolumeSpecName: "kube-api-access-6ntzp") pod "3e1c7427-d041-4489-89a1-e5b44ac89b30" (UID: "3e1c7427-d041-4489-89a1-e5b44ac89b30"). InnerVolumeSpecName "kube-api-access-6ntzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.421358 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e1c7427-d041-4489-89a1-e5b44ac89b30" (UID: "3e1c7427-d041-4489-89a1-e5b44ac89b30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.469800 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.469829 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ntzp\" (UniqueName: \"kubernetes.io/projected/3e1c7427-d041-4489-89a1-e5b44ac89b30-kube-api-access-6ntzp\") on node \"crc\" DevicePath \"\"" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.469838 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1c7427-d041-4489-89a1-e5b44ac89b30-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.779904 4741 generic.go:334] "Generic (PLEG): container finished" podID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerID="1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc" exitCode=0 Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.779958 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkkxf" event={"ID":"3e1c7427-d041-4489-89a1-e5b44ac89b30","Type":"ContainerDied","Data":"1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc"} Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.780002 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkkxf" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.780017 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkkxf" event={"ID":"3e1c7427-d041-4489-89a1-e5b44ac89b30","Type":"ContainerDied","Data":"a1b18cb197f4ade37664d97735e31b1c086e08847f62888197b29d9f578b2c03"} Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.780044 4741 scope.go:117] "RemoveContainer" containerID="1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.840657 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qkkxf"] Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.842027 4741 scope.go:117] "RemoveContainer" containerID="b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.860926 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qkkxf"] Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.865973 4741 scope.go:117] "RemoveContainer" containerID="6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.919794 4741 scope.go:117] "RemoveContainer" containerID="1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc" Sep 29 20:51:59 crc kubenswrapper[4741]: E0929 20:51:59.920468 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc\": container with ID starting with 1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc not found: ID does not exist" containerID="1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.920516 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc"} err="failed to get container status \"1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc\": rpc error: code = NotFound desc = could not find container \"1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc\": container with ID starting with 1357f3432fd1c7974a2c97a32b0b24496ad63ce3b550a843652f264e0657f3cc not found: ID does not exist" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.920547 4741 scope.go:117] "RemoveContainer" containerID="b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245" Sep 29 20:51:59 crc kubenswrapper[4741]: E0929 20:51:59.920812 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245\": container with ID starting with b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245 not found: ID does not exist" containerID="b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.920837 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245"} err="failed to get container status \"b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245\": rpc error: code = NotFound desc = could not find container \"b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245\": container with ID starting with b1e0525505921e64c25392d962b7927f5dfd09b3b6d0eda9e081233230511245 not found: ID does not exist" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.920855 4741 scope.go:117] "RemoveContainer" containerID="6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6" Sep 29 20:51:59 crc kubenswrapper[4741]: E0929 20:51:59.921059 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6\": container with ID starting with 6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6 not found: ID does not exist" containerID="6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6" Sep 29 20:51:59 crc kubenswrapper[4741]: I0929 20:51:59.921081 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6"} err="failed to get container status \"6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6\": rpc error: code = NotFound desc = could not find container \"6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6\": container with ID starting with 6521e0895132e1993f5c7b779b31c6e729819708efd9728f2fb23b98b93e33f6 not found: ID does not exist" Sep 29 20:52:01 crc kubenswrapper[4741]: I0929 20:52:01.147632 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" path="/var/lib/kubelet/pods/3e1c7427-d041-4489-89a1-e5b44ac89b30/volumes" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.579341 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-tldwl"] Sep 29 20:52:02 crc kubenswrapper[4741]: E0929 20:52:02.580076 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="registry-server" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.580090 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="registry-server" Sep 29 20:52:02 crc kubenswrapper[4741]: E0929 20:52:02.580127 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="extract-utilities" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.580134 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="extract-utilities" Sep 29 20:52:02 crc kubenswrapper[4741]: E0929 20:52:02.580157 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="extract-content" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.580164 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="extract-content" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.580356 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1c7427-d041-4489-89a1-e5b44ac89b30" containerName="registry-server" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.581165 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tldwl" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.601105 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-tldwl"] Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.651412 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whgc2\" (UniqueName: \"kubernetes.io/projected/73a732d9-cfa4-4cb0-b2a8-2643777b89fb-kube-api-access-whgc2\") pod \"aodh-db-create-tldwl\" (UID: \"73a732d9-cfa4-4cb0-b2a8-2643777b89fb\") " pod="openstack/aodh-db-create-tldwl" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.754034 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whgc2\" (UniqueName: \"kubernetes.io/projected/73a732d9-cfa4-4cb0-b2a8-2643777b89fb-kube-api-access-whgc2\") pod \"aodh-db-create-tldwl\" (UID: \"73a732d9-cfa4-4cb0-b2a8-2643777b89fb\") " pod="openstack/aodh-db-create-tldwl" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.770949 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whgc2\" (UniqueName: \"kubernetes.io/projected/73a732d9-cfa4-4cb0-b2a8-2643777b89fb-kube-api-access-whgc2\") pod \"aodh-db-create-tldwl\" (UID: \"73a732d9-cfa4-4cb0-b2a8-2643777b89fb\") " pod="openstack/aodh-db-create-tldwl" Sep 29 20:52:02 crc kubenswrapper[4741]: I0929 20:52:02.908468 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tldwl" Sep 29 20:52:03 crc kubenswrapper[4741]: W0929 20:52:03.401329 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73a732d9_cfa4_4cb0_b2a8_2643777b89fb.slice/crio-5c7d452a0591f89f661e50d0410a130d3d9ca924d7c9cd076d5413ba23574b46 WatchSource:0}: Error finding container 5c7d452a0591f89f661e50d0410a130d3d9ca924d7c9cd076d5413ba23574b46: Status 404 returned error can't find the container with id 5c7d452a0591f89f661e50d0410a130d3d9ca924d7c9cd076d5413ba23574b46 Sep 29 20:52:03 crc kubenswrapper[4741]: I0929 20:52:03.405323 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-tldwl"] Sep 29 20:52:03 crc kubenswrapper[4741]: I0929 20:52:03.834376 4741 generic.go:334] "Generic (PLEG): container finished" podID="73a732d9-cfa4-4cb0-b2a8-2643777b89fb" containerID="7c143506f7001f0f1b2cbf17d2e6db522814b3137b760d51e795abc005843ed2" exitCode=0 Sep 29 20:52:03 crc kubenswrapper[4741]: I0929 20:52:03.834484 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-tldwl" event={"ID":"73a732d9-cfa4-4cb0-b2a8-2643777b89fb","Type":"ContainerDied","Data":"7c143506f7001f0f1b2cbf17d2e6db522814b3137b760d51e795abc005843ed2"} Sep 29 20:52:03 crc kubenswrapper[4741]: I0929 20:52:03.834726 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-tldwl" event={"ID":"73a732d9-cfa4-4cb0-b2a8-2643777b89fb","Type":"ContainerStarted","Data":"5c7d452a0591f89f661e50d0410a130d3d9ca924d7c9cd076d5413ba23574b46"} Sep 29 20:52:05 crc kubenswrapper[4741]: I0929 20:52:05.286877 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tldwl" Sep 29 20:52:05 crc kubenswrapper[4741]: I0929 20:52:05.404040 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whgc2\" (UniqueName: \"kubernetes.io/projected/73a732d9-cfa4-4cb0-b2a8-2643777b89fb-kube-api-access-whgc2\") pod \"73a732d9-cfa4-4cb0-b2a8-2643777b89fb\" (UID: \"73a732d9-cfa4-4cb0-b2a8-2643777b89fb\") " Sep 29 20:52:05 crc kubenswrapper[4741]: I0929 20:52:05.410740 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73a732d9-cfa4-4cb0-b2a8-2643777b89fb-kube-api-access-whgc2" (OuterVolumeSpecName: "kube-api-access-whgc2") pod "73a732d9-cfa4-4cb0-b2a8-2643777b89fb" (UID: "73a732d9-cfa4-4cb0-b2a8-2643777b89fb"). InnerVolumeSpecName "kube-api-access-whgc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:52:05 crc kubenswrapper[4741]: I0929 20:52:05.506838 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whgc2\" (UniqueName: \"kubernetes.io/projected/73a732d9-cfa4-4cb0-b2a8-2643777b89fb-kube-api-access-whgc2\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:05 crc kubenswrapper[4741]: I0929 20:52:05.853221 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-tldwl" event={"ID":"73a732d9-cfa4-4cb0-b2a8-2643777b89fb","Type":"ContainerDied","Data":"5c7d452a0591f89f661e50d0410a130d3d9ca924d7c9cd076d5413ba23574b46"} Sep 29 20:52:05 crc kubenswrapper[4741]: I0929 20:52:05.853571 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c7d452a0591f89f661e50d0410a130d3d9ca924d7c9cd076d5413ba23574b46" Sep 29 20:52:05 crc kubenswrapper[4741]: I0929 20:52:05.853246 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tldwl" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.707007 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-a3d6-account-create-x59t8"] Sep 29 20:52:12 crc kubenswrapper[4741]: E0929 20:52:12.708613 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a732d9-cfa4-4cb0-b2a8-2643777b89fb" containerName="mariadb-database-create" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.708640 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a732d9-cfa4-4cb0-b2a8-2643777b89fb" containerName="mariadb-database-create" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.709252 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="73a732d9-cfa4-4cb0-b2a8-2643777b89fb" containerName="mariadb-database-create" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.710692 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a3d6-account-create-x59t8" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.713796 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.719250 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-a3d6-account-create-x59t8"] Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.765733 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swxr2\" (UniqueName: \"kubernetes.io/projected/8ed0b05b-c120-420f-907a-44aa6c04f7f8-kube-api-access-swxr2\") pod \"aodh-a3d6-account-create-x59t8\" (UID: \"8ed0b05b-c120-420f-907a-44aa6c04f7f8\") " pod="openstack/aodh-a3d6-account-create-x59t8" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.868781 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swxr2\" (UniqueName: \"kubernetes.io/projected/8ed0b05b-c120-420f-907a-44aa6c04f7f8-kube-api-access-swxr2\") pod \"aodh-a3d6-account-create-x59t8\" (UID: \"8ed0b05b-c120-420f-907a-44aa6c04f7f8\") " pod="openstack/aodh-a3d6-account-create-x59t8" Sep 29 20:52:12 crc kubenswrapper[4741]: I0929 20:52:12.923305 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swxr2\" (UniqueName: \"kubernetes.io/projected/8ed0b05b-c120-420f-907a-44aa6c04f7f8-kube-api-access-swxr2\") pod \"aodh-a3d6-account-create-x59t8\" (UID: \"8ed0b05b-c120-420f-907a-44aa6c04f7f8\") " pod="openstack/aodh-a3d6-account-create-x59t8" Sep 29 20:52:13 crc kubenswrapper[4741]: I0929 20:52:13.065455 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a3d6-account-create-x59t8" Sep 29 20:52:13 crc kubenswrapper[4741]: I0929 20:52:13.086821 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:52:13 crc kubenswrapper[4741]: E0929 20:52:13.087114 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:52:13 crc kubenswrapper[4741]: I0929 20:52:13.602145 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-a3d6-account-create-x59t8"] Sep 29 20:52:13 crc kubenswrapper[4741]: W0929 20:52:13.607425 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ed0b05b_c120_420f_907a_44aa6c04f7f8.slice/crio-023beb0b5db96ad5982ad7b19189a1f61de8d8ad51d156ae51ad46c95fb983ff WatchSource:0}: Error finding container 023beb0b5db96ad5982ad7b19189a1f61de8d8ad51d156ae51ad46c95fb983ff: Status 404 returned error can't find the container with id 023beb0b5db96ad5982ad7b19189a1f61de8d8ad51d156ae51ad46c95fb983ff Sep 29 20:52:13 crc kubenswrapper[4741]: I0929 20:52:13.947372 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-a3d6-account-create-x59t8" event={"ID":"8ed0b05b-c120-420f-907a-44aa6c04f7f8","Type":"ContainerStarted","Data":"6245bec3c4b5c97c308ec2dd89e7399399b87cb6d9de502362fffb0072c78a7e"} Sep 29 20:52:13 crc kubenswrapper[4741]: I0929 20:52:13.949007 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-a3d6-account-create-x59t8" event={"ID":"8ed0b05b-c120-420f-907a-44aa6c04f7f8","Type":"ContainerStarted","Data":"023beb0b5db96ad5982ad7b19189a1f61de8d8ad51d156ae51ad46c95fb983ff"} Sep 29 20:52:14 crc kubenswrapper[4741]: I0929 20:52:14.959786 4741 generic.go:334] "Generic (PLEG): container finished" podID="8ed0b05b-c120-420f-907a-44aa6c04f7f8" containerID="6245bec3c4b5c97c308ec2dd89e7399399b87cb6d9de502362fffb0072c78a7e" exitCode=0 Sep 29 20:52:14 crc kubenswrapper[4741]: I0929 20:52:14.960139 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-a3d6-account-create-x59t8" event={"ID":"8ed0b05b-c120-420f-907a-44aa6c04f7f8","Type":"ContainerDied","Data":"6245bec3c4b5c97c308ec2dd89e7399399b87cb6d9de502362fffb0072c78a7e"} Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.046954 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2lb6f"] Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.055500 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2lb6f"] Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.121536 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a43ba15-f4b7-45e1-a111-2aecebc65803" path="/var/lib/kubelet/pods/5a43ba15-f4b7-45e1-a111-2aecebc65803/volumes" Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.465041 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a3d6-account-create-x59t8" Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.528205 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swxr2\" (UniqueName: \"kubernetes.io/projected/8ed0b05b-c120-420f-907a-44aa6c04f7f8-kube-api-access-swxr2\") pod \"8ed0b05b-c120-420f-907a-44aa6c04f7f8\" (UID: \"8ed0b05b-c120-420f-907a-44aa6c04f7f8\") " Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.534697 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ed0b05b-c120-420f-907a-44aa6c04f7f8-kube-api-access-swxr2" (OuterVolumeSpecName: "kube-api-access-swxr2") pod "8ed0b05b-c120-420f-907a-44aa6c04f7f8" (UID: "8ed0b05b-c120-420f-907a-44aa6c04f7f8"). InnerVolumeSpecName "kube-api-access-swxr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.632696 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swxr2\" (UniqueName: \"kubernetes.io/projected/8ed0b05b-c120-420f-907a-44aa6c04f7f8-kube-api-access-swxr2\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.976064 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-a3d6-account-create-x59t8" event={"ID":"8ed0b05b-c120-420f-907a-44aa6c04f7f8","Type":"ContainerDied","Data":"023beb0b5db96ad5982ad7b19189a1f61de8d8ad51d156ae51ad46c95fb983ff"} Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.976589 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="023beb0b5db96ad5982ad7b19189a1f61de8d8ad51d156ae51ad46c95fb983ff" Sep 29 20:52:15 crc kubenswrapper[4741]: I0929 20:52:15.976110 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a3d6-account-create-x59t8" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.059051 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-jtjtg"] Sep 29 20:52:18 crc kubenswrapper[4741]: E0929 20:52:18.060370 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed0b05b-c120-420f-907a-44aa6c04f7f8" containerName="mariadb-account-create" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.060422 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed0b05b-c120-420f-907a-44aa6c04f7f8" containerName="mariadb-account-create" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.060952 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed0b05b-c120-420f-907a-44aa6c04f7f8" containerName="mariadb-account-create" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.062551 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.069938 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.070358 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.070603 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-x67km" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.074553 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jtjtg"] Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.096020 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-combined-ca-bundle\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.096110 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-scripts\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.096158 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75tx2\" (UniqueName: \"kubernetes.io/projected/09f18429-df86-4df8-aea7-7af91d6e6a2f-kube-api-access-75tx2\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.096257 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-config-data\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.198654 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-config-data\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.198726 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-combined-ca-bundle\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.198814 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-scripts\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.198848 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75tx2\" (UniqueName: \"kubernetes.io/projected/09f18429-df86-4df8-aea7-7af91d6e6a2f-kube-api-access-75tx2\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.208535 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-config-data\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.213659 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-scripts\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.214162 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-combined-ca-bundle\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.218263 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75tx2\" (UniqueName: \"kubernetes.io/projected/09f18429-df86-4df8-aea7-7af91d6e6a2f-kube-api-access-75tx2\") pod \"aodh-db-sync-jtjtg\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.408326 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:18 crc kubenswrapper[4741]: I0929 20:52:18.919517 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jtjtg"] Sep 29 20:52:19 crc kubenswrapper[4741]: I0929 20:52:19.014638 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jtjtg" event={"ID":"09f18429-df86-4df8-aea7-7af91d6e6a2f","Type":"ContainerStarted","Data":"fa3ed1f3eecbd357d7e66eabd7d1eecd97d7f4dabecb29566e16f495d173c3eb"} Sep 29 20:52:21 crc kubenswrapper[4741]: I0929 20:52:21.762206 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 29 20:52:23 crc kubenswrapper[4741]: I0929 20:52:23.059835 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jtjtg" event={"ID":"09f18429-df86-4df8-aea7-7af91d6e6a2f","Type":"ContainerStarted","Data":"c13ebe7d4bbd4ba511505a2c04b61b885477f90ca7b2510968cc9b0b6d683a66"} Sep 29 20:52:23 crc kubenswrapper[4741]: I0929 20:52:23.077680 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-jtjtg" podStartSLOduration=1.41018343 podStartE2EDuration="5.077662402s" podCreationTimestamp="2025-09-29 20:52:18 +0000 UTC" firstStartedPulling="2025-09-29 20:52:18.927370536 +0000 UTC m=+6180.575159868" lastFinishedPulling="2025-09-29 20:52:22.594849508 +0000 UTC m=+6184.242638840" observedRunningTime="2025-09-29 20:52:23.07410177 +0000 UTC m=+6184.721891102" watchObservedRunningTime="2025-09-29 20:52:23.077662402 +0000 UTC m=+6184.725451734" Sep 29 20:52:25 crc kubenswrapper[4741]: I0929 20:52:25.086233 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:52:25 crc kubenswrapper[4741]: E0929 20:52:25.086808 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:52:25 crc kubenswrapper[4741]: I0929 20:52:25.090341 4741 generic.go:334] "Generic (PLEG): container finished" podID="09f18429-df86-4df8-aea7-7af91d6e6a2f" containerID="c13ebe7d4bbd4ba511505a2c04b61b885477f90ca7b2510968cc9b0b6d683a66" exitCode=0 Sep 29 20:52:25 crc kubenswrapper[4741]: I0929 20:52:25.100604 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jtjtg" event={"ID":"09f18429-df86-4df8-aea7-7af91d6e6a2f","Type":"ContainerDied","Data":"c13ebe7d4bbd4ba511505a2c04b61b885477f90ca7b2510968cc9b0b6d683a66"} Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.487602 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.637485 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-scripts\") pod \"09f18429-df86-4df8-aea7-7af91d6e6a2f\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.637555 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-combined-ca-bundle\") pod \"09f18429-df86-4df8-aea7-7af91d6e6a2f\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.637640 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75tx2\" (UniqueName: \"kubernetes.io/projected/09f18429-df86-4df8-aea7-7af91d6e6a2f-kube-api-access-75tx2\") pod \"09f18429-df86-4df8-aea7-7af91d6e6a2f\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.637693 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-config-data\") pod \"09f18429-df86-4df8-aea7-7af91d6e6a2f\" (UID: \"09f18429-df86-4df8-aea7-7af91d6e6a2f\") " Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.652771 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f18429-df86-4df8-aea7-7af91d6e6a2f-kube-api-access-75tx2" (OuterVolumeSpecName: "kube-api-access-75tx2") pod "09f18429-df86-4df8-aea7-7af91d6e6a2f" (UID: "09f18429-df86-4df8-aea7-7af91d6e6a2f"). InnerVolumeSpecName "kube-api-access-75tx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.652770 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-scripts" (OuterVolumeSpecName: "scripts") pod "09f18429-df86-4df8-aea7-7af91d6e6a2f" (UID: "09f18429-df86-4df8-aea7-7af91d6e6a2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.668412 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-config-data" (OuterVolumeSpecName: "config-data") pod "09f18429-df86-4df8-aea7-7af91d6e6a2f" (UID: "09f18429-df86-4df8-aea7-7af91d6e6a2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.677552 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09f18429-df86-4df8-aea7-7af91d6e6a2f" (UID: "09f18429-df86-4df8-aea7-7af91d6e6a2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.740558 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75tx2\" (UniqueName: \"kubernetes.io/projected/09f18429-df86-4df8-aea7-7af91d6e6a2f-kube-api-access-75tx2\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.740760 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.740869 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:26 crc kubenswrapper[4741]: I0929 20:52:26.740947 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f18429-df86-4df8-aea7-7af91d6e6a2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.112423 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jtjtg" event={"ID":"09f18429-df86-4df8-aea7-7af91d6e6a2f","Type":"ContainerDied","Data":"fa3ed1f3eecbd357d7e66eabd7d1eecd97d7f4dabecb29566e16f495d173c3eb"} Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.112930 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa3ed1f3eecbd357d7e66eabd7d1eecd97d7f4dabecb29566e16f495d173c3eb" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.112998 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jtjtg" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.655420 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Sep 29 20:52:27 crc kubenswrapper[4741]: E0929 20:52:27.656073 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f18429-df86-4df8-aea7-7af91d6e6a2f" containerName="aodh-db-sync" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.656091 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f18429-df86-4df8-aea7-7af91d6e6a2f" containerName="aodh-db-sync" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.656357 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f18429-df86-4df8-aea7-7af91d6e6a2f" containerName="aodh-db-sync" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.659155 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.662044 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.662460 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-x67km" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.662685 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.673262 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.767306 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.767426 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-scripts\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.767596 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-config-data\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.767816 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dzwv\" (UniqueName: \"kubernetes.io/projected/7a4c3530-6b9e-47e4-a021-1abb149804b0-kube-api-access-7dzwv\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.870853 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-scripts\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.870989 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-config-data\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.871136 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dzwv\" (UniqueName: \"kubernetes.io/projected/7a4c3530-6b9e-47e4-a021-1abb149804b0-kube-api-access-7dzwv\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.871207 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.876376 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-scripts\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.877195 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.894113 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4c3530-6b9e-47e4-a021-1abb149804b0-config-data\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:27 crc kubenswrapper[4741]: I0929 20:52:27.894733 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dzwv\" (UniqueName: \"kubernetes.io/projected/7a4c3530-6b9e-47e4-a021-1abb149804b0-kube-api-access-7dzwv\") pod \"aodh-0\" (UID: \"7a4c3530-6b9e-47e4-a021-1abb149804b0\") " pod="openstack/aodh-0" Sep 29 20:52:28 crc kubenswrapper[4741]: I0929 20:52:28.031775 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Sep 29 20:52:28 crc kubenswrapper[4741]: I0929 20:52:28.520917 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Sep 29 20:52:29 crc kubenswrapper[4741]: I0929 20:52:29.147609 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7a4c3530-6b9e-47e4-a021-1abb149804b0","Type":"ContainerStarted","Data":"4f228d2d5c654cc8883d6f359db9b1a077ff1700759c0d25068dd1a9cec2558f"} Sep 29 20:52:29 crc kubenswrapper[4741]: I0929 20:52:29.438598 4741 scope.go:117] "RemoveContainer" containerID="6ad2d5b3090e1a80a515918e076f42c30f5497305ee46097585bb4506ae16479" Sep 29 20:52:29 crc kubenswrapper[4741]: I0929 20:52:29.489025 4741 scope.go:117] "RemoveContainer" containerID="486fb5c1b7e1d8ae7fdf980d6a6ce27ee1ee415f2e3560964ab4e4ad889197a2" Sep 29 20:52:29 crc kubenswrapper[4741]: I0929 20:52:29.547752 4741 scope.go:117] "RemoveContainer" containerID="9e372166fb4936206ea260fd5aeea6e1ba893242dd5e5122802e936310050c73" Sep 29 20:52:29 crc kubenswrapper[4741]: I0929 20:52:29.581135 4741 scope.go:117] "RemoveContainer" containerID="15bdcf1fd3f5ad55854ac6d22543b49505c17938181d65937cc7cc0c6a38d6d0" Sep 29 20:52:30 crc kubenswrapper[4741]: I0929 20:52:30.158530 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7a4c3530-6b9e-47e4-a021-1abb149804b0","Type":"ContainerStarted","Data":"19ba5a6ae85a3270d931d8349819fce143c909659de037b6994b190167152f0d"} Sep 29 20:52:30 crc kubenswrapper[4741]: I0929 20:52:30.252135 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:52:30 crc kubenswrapper[4741]: I0929 20:52:30.252556 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-central-agent" containerID="cri-o://1862b05a3f863ba3f54ed86d164e1cc1f8ea4f460f984abf5ef8f72ce7b83f8b" gracePeriod=30 Sep 29 20:52:30 crc kubenswrapper[4741]: I0929 20:52:30.252784 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-notification-agent" containerID="cri-o://c48f8b84d098762b1900b081283294ca4eaf14daf992db45cf8dfc9e10eebd4c" gracePeriod=30 Sep 29 20:52:30 crc kubenswrapper[4741]: I0929 20:52:30.252804 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="sg-core" containerID="cri-o://22dc64af77da9960a0f5093954cd0bbecc8a233de0e6c9b2763c029c7517285f" gracePeriod=30 Sep 29 20:52:30 crc kubenswrapper[4741]: I0929 20:52:30.252840 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="proxy-httpd" containerID="cri-o://0dbd4f9284d78e9080cf849ed6601ea67af721c050d29358648f8838411d35ae" gracePeriod=30 Sep 29 20:52:31 crc kubenswrapper[4741]: I0929 20:52:31.171832 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7a4c3530-6b9e-47e4-a021-1abb149804b0","Type":"ContainerStarted","Data":"fba8bb25326ca06636a1cf217106e6abdb19496acd680dc73b2a61977f5b1e71"} Sep 29 20:52:31 crc kubenswrapper[4741]: I0929 20:52:31.174750 4741 generic.go:334] "Generic (PLEG): container finished" podID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerID="0dbd4f9284d78e9080cf849ed6601ea67af721c050d29358648f8838411d35ae" exitCode=0 Sep 29 20:52:31 crc kubenswrapper[4741]: I0929 20:52:31.174783 4741 generic.go:334] "Generic (PLEG): container finished" podID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerID="22dc64af77da9960a0f5093954cd0bbecc8a233de0e6c9b2763c029c7517285f" exitCode=2 Sep 29 20:52:31 crc kubenswrapper[4741]: I0929 20:52:31.174796 4741 generic.go:334] "Generic (PLEG): container finished" podID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerID="1862b05a3f863ba3f54ed86d164e1cc1f8ea4f460f984abf5ef8f72ce7b83f8b" exitCode=0 Sep 29 20:52:31 crc kubenswrapper[4741]: I0929 20:52:31.174820 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerDied","Data":"0dbd4f9284d78e9080cf849ed6601ea67af721c050d29358648f8838411d35ae"} Sep 29 20:52:31 crc kubenswrapper[4741]: I0929 20:52:31.174863 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerDied","Data":"22dc64af77da9960a0f5093954cd0bbecc8a233de0e6c9b2763c029c7517285f"} Sep 29 20:52:31 crc kubenswrapper[4741]: I0929 20:52:31.174876 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerDied","Data":"1862b05a3f863ba3f54ed86d164e1cc1f8ea4f460f984abf5ef8f72ce7b83f8b"} Sep 29 20:52:32 crc kubenswrapper[4741]: I0929 20:52:32.188016 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7a4c3530-6b9e-47e4-a021-1abb149804b0","Type":"ContainerStarted","Data":"2a0a3330d69da38244cc9a4ee79f0e630ae606c5f63aa2fceec40f684c96eaa2"} Sep 29 20:52:34 crc kubenswrapper[4741]: I0929 20:52:34.211766 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7a4c3530-6b9e-47e4-a021-1abb149804b0","Type":"ContainerStarted","Data":"0217224480eafaeb1bf127c1482f0431c6942f25e53704d585e2c931276e3ef5"} Sep 29 20:52:34 crc kubenswrapper[4741]: I0929 20:52:34.231121 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.067159007 podStartE2EDuration="7.231104814s" podCreationTimestamp="2025-09-29 20:52:27 +0000 UTC" firstStartedPulling="2025-09-29 20:52:28.542424302 +0000 UTC m=+6190.190213634" lastFinishedPulling="2025-09-29 20:52:33.706370099 +0000 UTC m=+6195.354159441" observedRunningTime="2025-09-29 20:52:34.229337508 +0000 UTC m=+6195.877126840" watchObservedRunningTime="2025-09-29 20:52:34.231104814 +0000 UTC m=+6195.878894146" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.239061 4741 generic.go:334] "Generic (PLEG): container finished" podID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerID="c48f8b84d098762b1900b081283294ca4eaf14daf992db45cf8dfc9e10eebd4c" exitCode=0 Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.239589 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerDied","Data":"c48f8b84d098762b1900b081283294ca4eaf14daf992db45cf8dfc9e10eebd4c"} Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.616845 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.780364 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-config-data\") pod \"27f1d9d0-0481-4f24-840b-4040cbf232e5\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.780475 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-log-httpd\") pod \"27f1d9d0-0481-4f24-840b-4040cbf232e5\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.780554 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kdbb\" (UniqueName: \"kubernetes.io/projected/27f1d9d0-0481-4f24-840b-4040cbf232e5-kube-api-access-2kdbb\") pod \"27f1d9d0-0481-4f24-840b-4040cbf232e5\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.780611 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-sg-core-conf-yaml\") pod \"27f1d9d0-0481-4f24-840b-4040cbf232e5\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.780655 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-run-httpd\") pod \"27f1d9d0-0481-4f24-840b-4040cbf232e5\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.781107 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "27f1d9d0-0481-4f24-840b-4040cbf232e5" (UID: "27f1d9d0-0481-4f24-840b-4040cbf232e5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.781126 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "27f1d9d0-0481-4f24-840b-4040cbf232e5" (UID: "27f1d9d0-0481-4f24-840b-4040cbf232e5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.781380 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-scripts\") pod \"27f1d9d0-0481-4f24-840b-4040cbf232e5\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.781447 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-combined-ca-bundle\") pod \"27f1d9d0-0481-4f24-840b-4040cbf232e5\" (UID: \"27f1d9d0-0481-4f24-840b-4040cbf232e5\") " Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.782003 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.782027 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27f1d9d0-0481-4f24-840b-4040cbf232e5-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.788590 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-scripts" (OuterVolumeSpecName: "scripts") pod "27f1d9d0-0481-4f24-840b-4040cbf232e5" (UID: "27f1d9d0-0481-4f24-840b-4040cbf232e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.788684 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27f1d9d0-0481-4f24-840b-4040cbf232e5-kube-api-access-2kdbb" (OuterVolumeSpecName: "kube-api-access-2kdbb") pod "27f1d9d0-0481-4f24-840b-4040cbf232e5" (UID: "27f1d9d0-0481-4f24-840b-4040cbf232e5"). InnerVolumeSpecName "kube-api-access-2kdbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.812024 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "27f1d9d0-0481-4f24-840b-4040cbf232e5" (UID: "27f1d9d0-0481-4f24-840b-4040cbf232e5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.851328 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27f1d9d0-0481-4f24-840b-4040cbf232e5" (UID: "27f1d9d0-0481-4f24-840b-4040cbf232e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.883837 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.883869 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.883884 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kdbb\" (UniqueName: \"kubernetes.io/projected/27f1d9d0-0481-4f24-840b-4040cbf232e5-kube-api-access-2kdbb\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.883896 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.905377 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-config-data" (OuterVolumeSpecName: "config-data") pod "27f1d9d0-0481-4f24-840b-4040cbf232e5" (UID: "27f1d9d0-0481-4f24-840b-4040cbf232e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:52:36 crc kubenswrapper[4741]: I0929 20:52:36.985583 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f1d9d0-0481-4f24-840b-4040cbf232e5-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.086651 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:52:37 crc kubenswrapper[4741]: E0929 20:52:37.087011 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.251562 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27f1d9d0-0481-4f24-840b-4040cbf232e5","Type":"ContainerDied","Data":"a45d5ce8f638953112136c50f1c32881ba8eb9da434c7cd327f31d814809355e"} Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.251619 4741 scope.go:117] "RemoveContainer" containerID="0dbd4f9284d78e9080cf849ed6601ea67af721c050d29358648f8838411d35ae" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.252622 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.333061 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.345191 4741 scope.go:117] "RemoveContainer" containerID="22dc64af77da9960a0f5093954cd0bbecc8a233de0e6c9b2763c029c7517285f" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.348717 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.362840 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:52:37 crc kubenswrapper[4741]: E0929 20:52:37.363542 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-central-agent" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.363574 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-central-agent" Sep 29 20:52:37 crc kubenswrapper[4741]: E0929 20:52:37.363613 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="proxy-httpd" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.363626 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="proxy-httpd" Sep 29 20:52:37 crc kubenswrapper[4741]: E0929 20:52:37.363651 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="sg-core" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.363662 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="sg-core" Sep 29 20:52:37 crc kubenswrapper[4741]: E0929 20:52:37.363692 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-notification-agent" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.363704 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-notification-agent" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.364102 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-notification-agent" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.364151 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="sg-core" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.364173 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="ceilometer-central-agent" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.364197 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" containerName="proxy-httpd" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.367763 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.370505 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.370539 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.375747 4741 scope.go:117] "RemoveContainer" containerID="c48f8b84d098762b1900b081283294ca4eaf14daf992db45cf8dfc9e10eebd4c" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.377732 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.416052 4741 scope.go:117] "RemoveContainer" containerID="1862b05a3f863ba3f54ed86d164e1cc1f8ea4f460f984abf5ef8f72ce7b83f8b" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.498911 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-run-httpd\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.498960 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-config-data\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.498994 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.499314 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.499513 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-scripts\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.499569 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-log-httpd\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.499791 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmt2p\" (UniqueName: \"kubernetes.io/projected/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-kube-api-access-pmt2p\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.601774 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.601866 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-scripts\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.601892 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-log-httpd\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.601934 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmt2p\" (UniqueName: \"kubernetes.io/projected/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-kube-api-access-pmt2p\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.601975 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-run-httpd\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.602000 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-config-data\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.602024 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.602887 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-run-httpd\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.603555 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-log-httpd\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.605462 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.605925 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-config-data\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.606239 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.619865 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmt2p\" (UniqueName: \"kubernetes.io/projected/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-kube-api-access-pmt2p\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.620388 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-scripts\") pod \"ceilometer-0\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " pod="openstack/ceilometer-0" Sep 29 20:52:37 crc kubenswrapper[4741]: I0929 20:52:37.719546 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:52:38 crc kubenswrapper[4741]: I0929 20:52:38.213615 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:52:38 crc kubenswrapper[4741]: I0929 20:52:38.262693 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerStarted","Data":"957341d5211bdaa10972588063c162e0aecf8719bb4f78d53f518d596628683b"} Sep 29 20:52:39 crc kubenswrapper[4741]: I0929 20:52:39.101678 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27f1d9d0-0481-4f24-840b-4040cbf232e5" path="/var/lib/kubelet/pods/27f1d9d0-0481-4f24-840b-4040cbf232e5/volumes" Sep 29 20:52:39 crc kubenswrapper[4741]: I0929 20:52:39.274355 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerStarted","Data":"da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616"} Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.294112 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerStarted","Data":"b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218"} Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.650695 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-lnkrj"] Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.652170 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lnkrj" Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.670752 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-lnkrj"] Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.768758 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js4m4\" (UniqueName: \"kubernetes.io/projected/353fdc30-16fe-4745-abbd-258072064666-kube-api-access-js4m4\") pod \"manila-db-create-lnkrj\" (UID: \"353fdc30-16fe-4745-abbd-258072064666\") " pod="openstack/manila-db-create-lnkrj" Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.871499 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js4m4\" (UniqueName: \"kubernetes.io/projected/353fdc30-16fe-4745-abbd-258072064666-kube-api-access-js4m4\") pod \"manila-db-create-lnkrj\" (UID: \"353fdc30-16fe-4745-abbd-258072064666\") " pod="openstack/manila-db-create-lnkrj" Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.894047 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js4m4\" (UniqueName: \"kubernetes.io/projected/353fdc30-16fe-4745-abbd-258072064666-kube-api-access-js4m4\") pod \"manila-db-create-lnkrj\" (UID: \"353fdc30-16fe-4745-abbd-258072064666\") " pod="openstack/manila-db-create-lnkrj" Sep 29 20:52:40 crc kubenswrapper[4741]: I0929 20:52:40.973319 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lnkrj" Sep 29 20:52:41 crc kubenswrapper[4741]: I0929 20:52:41.312618 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerStarted","Data":"e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e"} Sep 29 20:52:41 crc kubenswrapper[4741]: I0929 20:52:41.480227 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-lnkrj"] Sep 29 20:52:41 crc kubenswrapper[4741]: E0929 20:52:41.970766 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod353fdc30_16fe_4745_abbd_258072064666.slice/crio-f66e81ec2e7929dee53f5032ae101d7d8e8a32ab1b65dde91b9aa5d2b9aeffdd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod353fdc30_16fe_4745_abbd_258072064666.slice/crio-conmon-f66e81ec2e7929dee53f5032ae101d7d8e8a32ab1b65dde91b9aa5d2b9aeffdd.scope\": RecentStats: unable to find data in memory cache]" Sep 29 20:52:42 crc kubenswrapper[4741]: I0929 20:52:42.322042 4741 generic.go:334] "Generic (PLEG): container finished" podID="353fdc30-16fe-4745-abbd-258072064666" containerID="f66e81ec2e7929dee53f5032ae101d7d8e8a32ab1b65dde91b9aa5d2b9aeffdd" exitCode=0 Sep 29 20:52:42 crc kubenswrapper[4741]: I0929 20:52:42.322122 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-lnkrj" event={"ID":"353fdc30-16fe-4745-abbd-258072064666","Type":"ContainerDied","Data":"f66e81ec2e7929dee53f5032ae101d7d8e8a32ab1b65dde91b9aa5d2b9aeffdd"} Sep 29 20:52:42 crc kubenswrapper[4741]: I0929 20:52:42.322155 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-lnkrj" event={"ID":"353fdc30-16fe-4745-abbd-258072064666","Type":"ContainerStarted","Data":"5465a8680507ce7648cf99b4ca2a1188d88e203fc6cc760512914dda503dd4f9"} Sep 29 20:52:42 crc kubenswrapper[4741]: I0929 20:52:42.325166 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerStarted","Data":"31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1"} Sep 29 20:52:42 crc kubenswrapper[4741]: I0929 20:52:42.325443 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 20:52:42 crc kubenswrapper[4741]: I0929 20:52:42.353834 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.193761722 podStartE2EDuration="5.353821755s" podCreationTimestamp="2025-09-29 20:52:37 +0000 UTC" firstStartedPulling="2025-09-29 20:52:38.202104744 +0000 UTC m=+6199.849894076" lastFinishedPulling="2025-09-29 20:52:41.362164777 +0000 UTC m=+6203.009954109" observedRunningTime="2025-09-29 20:52:42.350621964 +0000 UTC m=+6203.998411296" watchObservedRunningTime="2025-09-29 20:52:42.353821755 +0000 UTC m=+6204.001611087" Sep 29 20:52:43 crc kubenswrapper[4741]: I0929 20:52:43.772214 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lnkrj" Sep 29 20:52:43 crc kubenswrapper[4741]: I0929 20:52:43.840928 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js4m4\" (UniqueName: \"kubernetes.io/projected/353fdc30-16fe-4745-abbd-258072064666-kube-api-access-js4m4\") pod \"353fdc30-16fe-4745-abbd-258072064666\" (UID: \"353fdc30-16fe-4745-abbd-258072064666\") " Sep 29 20:52:43 crc kubenswrapper[4741]: I0929 20:52:43.847603 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/353fdc30-16fe-4745-abbd-258072064666-kube-api-access-js4m4" (OuterVolumeSpecName: "kube-api-access-js4m4") pod "353fdc30-16fe-4745-abbd-258072064666" (UID: "353fdc30-16fe-4745-abbd-258072064666"). InnerVolumeSpecName "kube-api-access-js4m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:52:43 crc kubenswrapper[4741]: I0929 20:52:43.943661 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js4m4\" (UniqueName: \"kubernetes.io/projected/353fdc30-16fe-4745-abbd-258072064666-kube-api-access-js4m4\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:44 crc kubenswrapper[4741]: I0929 20:52:44.343979 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-lnkrj" event={"ID":"353fdc30-16fe-4745-abbd-258072064666","Type":"ContainerDied","Data":"5465a8680507ce7648cf99b4ca2a1188d88e203fc6cc760512914dda503dd4f9"} Sep 29 20:52:44 crc kubenswrapper[4741]: I0929 20:52:44.344021 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5465a8680507ce7648cf99b4ca2a1188d88e203fc6cc760512914dda503dd4f9" Sep 29 20:52:44 crc kubenswrapper[4741]: I0929 20:52:44.344113 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lnkrj" Sep 29 20:52:49 crc kubenswrapper[4741]: I0929 20:52:49.103987 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:52:49 crc kubenswrapper[4741]: E0929 20:52:49.105320 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:52:50 crc kubenswrapper[4741]: I0929 20:52:50.792928 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-4428-account-create-p8bhg"] Sep 29 20:52:50 crc kubenswrapper[4741]: E0929 20:52:50.793983 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353fdc30-16fe-4745-abbd-258072064666" containerName="mariadb-database-create" Sep 29 20:52:50 crc kubenswrapper[4741]: I0929 20:52:50.794012 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="353fdc30-16fe-4745-abbd-258072064666" containerName="mariadb-database-create" Sep 29 20:52:50 crc kubenswrapper[4741]: I0929 20:52:50.794470 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="353fdc30-16fe-4745-abbd-258072064666" containerName="mariadb-database-create" Sep 29 20:52:50 crc kubenswrapper[4741]: I0929 20:52:50.795519 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-4428-account-create-p8bhg" Sep 29 20:52:50 crc kubenswrapper[4741]: I0929 20:52:50.797714 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Sep 29 20:52:50 crc kubenswrapper[4741]: I0929 20:52:50.802671 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-4428-account-create-p8bhg"] Sep 29 20:52:50 crc kubenswrapper[4741]: I0929 20:52:50.901603 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f789b\" (UniqueName: \"kubernetes.io/projected/3bc712f3-183e-4421-9238-e48591525259-kube-api-access-f789b\") pod \"manila-4428-account-create-p8bhg\" (UID: \"3bc712f3-183e-4421-9238-e48591525259\") " pod="openstack/manila-4428-account-create-p8bhg" Sep 29 20:52:51 crc kubenswrapper[4741]: I0929 20:52:51.003821 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f789b\" (UniqueName: \"kubernetes.io/projected/3bc712f3-183e-4421-9238-e48591525259-kube-api-access-f789b\") pod \"manila-4428-account-create-p8bhg\" (UID: \"3bc712f3-183e-4421-9238-e48591525259\") " pod="openstack/manila-4428-account-create-p8bhg" Sep 29 20:52:51 crc kubenswrapper[4741]: I0929 20:52:51.022438 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f789b\" (UniqueName: \"kubernetes.io/projected/3bc712f3-183e-4421-9238-e48591525259-kube-api-access-f789b\") pod \"manila-4428-account-create-p8bhg\" (UID: \"3bc712f3-183e-4421-9238-e48591525259\") " pod="openstack/manila-4428-account-create-p8bhg" Sep 29 20:52:51 crc kubenswrapper[4741]: I0929 20:52:51.130138 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-4428-account-create-p8bhg" Sep 29 20:52:51 crc kubenswrapper[4741]: I0929 20:52:51.658307 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-4428-account-create-p8bhg"] Sep 29 20:52:52 crc kubenswrapper[4741]: I0929 20:52:52.417889 4741 generic.go:334] "Generic (PLEG): container finished" podID="3bc712f3-183e-4421-9238-e48591525259" containerID="ee03b83dd6a02bacfbb25bdd8a3ecc5aad2e652eb3d421c96398e4e4f58a3705" exitCode=0 Sep 29 20:52:52 crc kubenswrapper[4741]: I0929 20:52:52.417941 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-4428-account-create-p8bhg" event={"ID":"3bc712f3-183e-4421-9238-e48591525259","Type":"ContainerDied","Data":"ee03b83dd6a02bacfbb25bdd8a3ecc5aad2e652eb3d421c96398e4e4f58a3705"} Sep 29 20:52:52 crc kubenswrapper[4741]: I0929 20:52:52.418284 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-4428-account-create-p8bhg" event={"ID":"3bc712f3-183e-4421-9238-e48591525259","Type":"ContainerStarted","Data":"0e2be234c4f1785e14e12b8c67b869c9e98cb5d45981fd537134c97bcd9c6e8b"} Sep 29 20:52:53 crc kubenswrapper[4741]: I0929 20:52:53.897075 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-4428-account-create-p8bhg" Sep 29 20:52:53 crc kubenswrapper[4741]: I0929 20:52:53.967172 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f789b\" (UniqueName: \"kubernetes.io/projected/3bc712f3-183e-4421-9238-e48591525259-kube-api-access-f789b\") pod \"3bc712f3-183e-4421-9238-e48591525259\" (UID: \"3bc712f3-183e-4421-9238-e48591525259\") " Sep 29 20:52:53 crc kubenswrapper[4741]: I0929 20:52:53.972566 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc712f3-183e-4421-9238-e48591525259-kube-api-access-f789b" (OuterVolumeSpecName: "kube-api-access-f789b") pod "3bc712f3-183e-4421-9238-e48591525259" (UID: "3bc712f3-183e-4421-9238-e48591525259"). InnerVolumeSpecName "kube-api-access-f789b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:52:54 crc kubenswrapper[4741]: I0929 20:52:54.070149 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f789b\" (UniqueName: \"kubernetes.io/projected/3bc712f3-183e-4421-9238-e48591525259-kube-api-access-f789b\") on node \"crc\" DevicePath \"\"" Sep 29 20:52:54 crc kubenswrapper[4741]: I0929 20:52:54.438920 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-4428-account-create-p8bhg" event={"ID":"3bc712f3-183e-4421-9238-e48591525259","Type":"ContainerDied","Data":"0e2be234c4f1785e14e12b8c67b869c9e98cb5d45981fd537134c97bcd9c6e8b"} Sep 29 20:52:54 crc kubenswrapper[4741]: I0929 20:52:54.439297 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e2be234c4f1785e14e12b8c67b869c9e98cb5d45981fd537134c97bcd9c6e8b" Sep 29 20:52:54 crc kubenswrapper[4741]: I0929 20:52:54.439029 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-4428-account-create-p8bhg" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.166751 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-tklfv"] Sep 29 20:52:56 crc kubenswrapper[4741]: E0929 20:52:56.168583 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc712f3-183e-4421-9238-e48591525259" containerName="mariadb-account-create" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.168602 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc712f3-183e-4421-9238-e48591525259" containerName="mariadb-account-create" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.169655 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc712f3-183e-4421-9238-e48591525259" containerName="mariadb-account-create" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.171597 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.173787 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-xwq4f" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.174163 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.177998 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-tklfv"] Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.227853 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-config-data\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.228110 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-job-config-data\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.228192 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfcjn\" (UniqueName: \"kubernetes.io/projected/51a5e675-eeec-4197-b435-fb03bd19662b-kube-api-access-wfcjn\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.228515 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-combined-ca-bundle\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.332193 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-job-config-data\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.332265 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfcjn\" (UniqueName: \"kubernetes.io/projected/51a5e675-eeec-4197-b435-fb03bd19662b-kube-api-access-wfcjn\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.332314 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-combined-ca-bundle\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.332577 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-config-data\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.337944 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-combined-ca-bundle\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.339463 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-config-data\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.340009 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-job-config-data\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.351916 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfcjn\" (UniqueName: \"kubernetes.io/projected/51a5e675-eeec-4197-b435-fb03bd19662b-kube-api-access-wfcjn\") pod \"manila-db-sync-tklfv\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:56 crc kubenswrapper[4741]: I0929 20:52:56.546550 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-tklfv" Sep 29 20:52:57 crc kubenswrapper[4741]: I0929 20:52:57.388748 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-tklfv"] Sep 29 20:52:57 crc kubenswrapper[4741]: I0929 20:52:57.468456 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-tklfv" event={"ID":"51a5e675-eeec-4197-b435-fb03bd19662b","Type":"ContainerStarted","Data":"199b420dde8128082b40b10a444191eccd7712bbe91842eb6eec5d4af17b4e31"} Sep 29 20:52:58 crc kubenswrapper[4741]: I0929 20:52:58.055749 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-dc4f2"] Sep 29 20:52:58 crc kubenswrapper[4741]: I0929 20:52:58.064746 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-dc4f2"] Sep 29 20:52:59 crc kubenswrapper[4741]: I0929 20:52:59.100239 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e93df880-9269-48e4-9815-bbd513579722" path="/var/lib/kubelet/pods/e93df880-9269-48e4-9815-bbd513579722/volumes" Sep 29 20:53:00 crc kubenswrapper[4741]: I0929 20:53:00.086963 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:53:00 crc kubenswrapper[4741]: E0929 20:53:00.087520 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:53:02 crc kubenswrapper[4741]: I0929 20:53:02.535846 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-tklfv" event={"ID":"51a5e675-eeec-4197-b435-fb03bd19662b","Type":"ContainerStarted","Data":"8e44993d8028dab55ed0ee5bdec553a980a1503d05a17c10a7b1ec63aff26cb8"} Sep 29 20:53:02 crc kubenswrapper[4741]: I0929 20:53:02.566774 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-tklfv" podStartSLOduration=2.5144905189999998 podStartE2EDuration="6.566754978s" podCreationTimestamp="2025-09-29 20:52:56 +0000 UTC" firstStartedPulling="2025-09-29 20:52:57.38721694 +0000 UTC m=+6219.035006272" lastFinishedPulling="2025-09-29 20:53:01.439481399 +0000 UTC m=+6223.087270731" observedRunningTime="2025-09-29 20:53:02.554938515 +0000 UTC m=+6224.202727887" watchObservedRunningTime="2025-09-29 20:53:02.566754978 +0000 UTC m=+6224.214544310" Sep 29 20:53:04 crc kubenswrapper[4741]: I0929 20:53:04.560684 4741 generic.go:334] "Generic (PLEG): container finished" podID="51a5e675-eeec-4197-b435-fb03bd19662b" containerID="8e44993d8028dab55ed0ee5bdec553a980a1503d05a17c10a7b1ec63aff26cb8" exitCode=0 Sep 29 20:53:04 crc kubenswrapper[4741]: I0929 20:53:04.561327 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-tklfv" event={"ID":"51a5e675-eeec-4197-b435-fb03bd19662b","Type":"ContainerDied","Data":"8e44993d8028dab55ed0ee5bdec553a980a1503d05a17c10a7b1ec63aff26cb8"} Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.127789 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-tklfv" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.166461 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-job-config-data\") pod \"51a5e675-eeec-4197-b435-fb03bd19662b\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.166862 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-config-data\") pod \"51a5e675-eeec-4197-b435-fb03bd19662b\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.167079 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-combined-ca-bundle\") pod \"51a5e675-eeec-4197-b435-fb03bd19662b\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.167109 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfcjn\" (UniqueName: \"kubernetes.io/projected/51a5e675-eeec-4197-b435-fb03bd19662b-kube-api-access-wfcjn\") pod \"51a5e675-eeec-4197-b435-fb03bd19662b\" (UID: \"51a5e675-eeec-4197-b435-fb03bd19662b\") " Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.175381 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a5e675-eeec-4197-b435-fb03bd19662b-kube-api-access-wfcjn" (OuterVolumeSpecName: "kube-api-access-wfcjn") pod "51a5e675-eeec-4197-b435-fb03bd19662b" (UID: "51a5e675-eeec-4197-b435-fb03bd19662b"). InnerVolumeSpecName "kube-api-access-wfcjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.177261 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-config-data" (OuterVolumeSpecName: "config-data") pod "51a5e675-eeec-4197-b435-fb03bd19662b" (UID: "51a5e675-eeec-4197-b435-fb03bd19662b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.177961 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "51a5e675-eeec-4197-b435-fb03bd19662b" (UID: "51a5e675-eeec-4197-b435-fb03bd19662b"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.201523 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51a5e675-eeec-4197-b435-fb03bd19662b" (UID: "51a5e675-eeec-4197-b435-fb03bd19662b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.270792 4741 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-job-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.270867 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.270887 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a5e675-eeec-4197-b435-fb03bd19662b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.270908 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfcjn\" (UniqueName: \"kubernetes.io/projected/51a5e675-eeec-4197-b435-fb03bd19662b-kube-api-access-wfcjn\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.583980 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-tklfv" event={"ID":"51a5e675-eeec-4197-b435-fb03bd19662b","Type":"ContainerDied","Data":"199b420dde8128082b40b10a444191eccd7712bbe91842eb6eec5d4af17b4e31"} Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.584271 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="199b420dde8128082b40b10a444191eccd7712bbe91842eb6eec5d4af17b4e31" Sep 29 20:53:06 crc kubenswrapper[4741]: I0929 20:53:06.584464 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-tklfv" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.065679 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Sep 29 20:53:07 crc kubenswrapper[4741]: E0929 20:53:07.066152 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a5e675-eeec-4197-b435-fb03bd19662b" containerName="manila-db-sync" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.066168 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a5e675-eeec-4197-b435-fb03bd19662b" containerName="manila-db-sync" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.083479 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a5e675-eeec-4197-b435-fb03bd19662b" containerName="manila-db-sync" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.102876 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.111347 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.111447 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.111789 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-xwq4f" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.127850 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.152921 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.165286 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.169280 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.194059 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.194143 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-config-data\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.194197 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.194284 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-scripts\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.194306 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkprb\" (UniqueName: \"kubernetes.io/projected/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-kube-api-access-xkprb\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.194343 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.195629 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.227245 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.246710 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-81ca-account-create-jhlwq"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.275817 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-81ca-account-create-jhlwq"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.297985 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298050 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdt45\" (UniqueName: \"kubernetes.io/projected/a28a5999-6348-4d83-9605-16292704fb39-kube-api-access-rdt45\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298078 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-config-data\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298099 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298120 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-scripts\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298168 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-scripts\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298190 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkprb\" (UniqueName: \"kubernetes.io/projected/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-kube-api-access-xkprb\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298221 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298263 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298301 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a28a5999-6348-4d83-9605-16292704fb39-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298315 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a28a5999-6348-4d83-9605-16292704fb39-ceph\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298336 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298354 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-config-data\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.298377 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a28a5999-6348-4d83-9605-16292704fb39-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.299610 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.303257 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.304284 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-scripts\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.304376 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-config-data\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.315803 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.320366 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkprb\" (UniqueName: \"kubernetes.io/projected/cf1826cd-d5f3-4ed1-ad4e-9646edde3a65-kube-api-access-xkprb\") pod \"manila-scheduler-0\" (UID: \"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65\") " pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.330643 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-797bf58c57-x7s92"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.333025 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.369298 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bf58c57-x7s92"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.379549 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.381652 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.384667 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.392664 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400582 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-dns-svc\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400626 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jdms\" (UniqueName: \"kubernetes.io/projected/639e8429-8817-42f6-bdea-1a96545e6640-kube-api-access-8jdms\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400650 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-sb\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400698 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-nb\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400729 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a28a5999-6348-4d83-9605-16292704fb39-ceph\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400749 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a28a5999-6348-4d83-9605-16292704fb39-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400775 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400805 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a28a5999-6348-4d83-9605-16292704fb39-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400845 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdt45\" (UniqueName: \"kubernetes.io/projected/a28a5999-6348-4d83-9605-16292704fb39-kube-api-access-rdt45\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400863 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-config-data\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400882 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-config\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400900 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.400918 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-scripts\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.402495 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a28a5999-6348-4d83-9605-16292704fb39-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.405447 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-scripts\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.409279 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a28a5999-6348-4d83-9605-16292704fb39-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.409813 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a28a5999-6348-4d83-9605-16292704fb39-ceph\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.411878 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.420659 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.438382 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28a5999-6348-4d83-9605-16292704fb39-config-data\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.445301 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdt45\" (UniqueName: \"kubernetes.io/projected/a28a5999-6348-4d83-9605-16292704fb39-kube-api-access-rdt45\") pod \"manila-share-share1-0\" (UID: \"a28a5999-6348-4d83-9605-16292704fb39\") " pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.463012 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504790 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-scripts\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504842 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-logs\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504890 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-dns-svc\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504911 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jdms\" (UniqueName: \"kubernetes.io/projected/639e8429-8817-42f6-bdea-1a96545e6640-kube-api-access-8jdms\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504933 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-sb\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504949 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-etc-machine-id\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504974 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.504988 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d6b6\" (UniqueName: \"kubernetes.io/projected/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-kube-api-access-7d6b6\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.505025 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-nb\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.505076 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-config-data\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.505116 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-config\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.505147 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-config-data-custom\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.511517 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-sb\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.511927 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-dns-svc\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.512233 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-nb\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.513008 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.514066 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-config\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.559187 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jdms\" (UniqueName: \"kubernetes.io/projected/639e8429-8817-42f6-bdea-1a96545e6640-kube-api-access-8jdms\") pod \"dnsmasq-dns-797bf58c57-x7s92\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.608447 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-etc-machine-id\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.609944 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.609991 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d6b6\" (UniqueName: \"kubernetes.io/projected/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-kube-api-access-7d6b6\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.610506 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-config-data\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.610785 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-config-data-custom\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.611052 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-scripts\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.611122 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-logs\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.611887 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-logs\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.609640 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-etc-machine-id\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.617893 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-scripts\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.618451 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.621308 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-config-data\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.631318 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-config-data-custom\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.638889 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.645211 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d6b6\" (UniqueName: \"kubernetes.io/projected/9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6-kube-api-access-7d6b6\") pod \"manila-api-0\" (UID: \"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6\") " pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.657055 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Sep 29 20:53:07 crc kubenswrapper[4741]: I0929 20:53:07.726785 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.301655 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.405707 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Sep 29 20:53:08 crc kubenswrapper[4741]: W0929 20:53:08.446055 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda28a5999_6348_4d83_9605_16292704fb39.slice/crio-3d0b409ae1a42ab4ff008370fb0a8df54e65e8c813d4989253591cfb94a9b044 WatchSource:0}: Error finding container 3d0b409ae1a42ab4ff008370fb0a8df54e65e8c813d4989253591cfb94a9b044: Status 404 returned error can't find the container with id 3d0b409ae1a42ab4ff008370fb0a8df54e65e8c813d4989253591cfb94a9b044 Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.456861 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bf58c57-x7s92"] Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.545791 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.630537 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65","Type":"ContainerStarted","Data":"864ed527fafc7235ce70817113bebcb75ba893c2e03ac7c9620629aae6f87d16"} Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.632865 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" event={"ID":"639e8429-8817-42f6-bdea-1a96545e6640","Type":"ContainerStarted","Data":"e711b0df64968ad06334c38425a9ae66674de6146349d4b09aebcccdf145e917"} Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.638499 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6","Type":"ContainerStarted","Data":"466b56ba2d39203ff144204cdb2f1cc11fc41cc95da6addf5f9e2e4e36772c1c"} Sep 29 20:53:08 crc kubenswrapper[4741]: I0929 20:53:08.640062 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a28a5999-6348-4d83-9605-16292704fb39","Type":"ContainerStarted","Data":"3d0b409ae1a42ab4ff008370fb0a8df54e65e8c813d4989253591cfb94a9b044"} Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.103165 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f3900c-8af5-484e-b0f3-3a60fdcc33e8" path="/var/lib/kubelet/pods/e4f3900c-8af5-484e-b0f3-3a60fdcc33e8/volumes" Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.681185 4741 generic.go:334] "Generic (PLEG): container finished" podID="639e8429-8817-42f6-bdea-1a96545e6640" containerID="2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c" exitCode=0 Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.681550 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" event={"ID":"639e8429-8817-42f6-bdea-1a96545e6640","Type":"ContainerDied","Data":"2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c"} Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.695873 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6","Type":"ContainerStarted","Data":"b5605edc164edfd9943ebf4cf60dd223d9b4a08a1519b7e4c7d18a83059cf3fe"} Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.695922 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6","Type":"ContainerStarted","Data":"46ff63fadfad9bb6b76392a190094c7a34844212c0cac24902b6c74b37b03c39"} Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.696843 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.722273 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65","Type":"ContainerStarted","Data":"05dd04847ea3374a14370231236f6bedca12dbade8fe54dcc609c8bd325da7e4"} Sep 29 20:53:09 crc kubenswrapper[4741]: I0929 20:53:09.751841 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=2.751822474 podStartE2EDuration="2.751822474s" podCreationTimestamp="2025-09-29 20:53:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:53:09.734829409 +0000 UTC m=+6231.382618761" watchObservedRunningTime="2025-09-29 20:53:09.751822474 +0000 UTC m=+6231.399611806" Sep 29 20:53:10 crc kubenswrapper[4741]: I0929 20:53:10.736179 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" event={"ID":"639e8429-8817-42f6-bdea-1a96545e6640","Type":"ContainerStarted","Data":"608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259"} Sep 29 20:53:10 crc kubenswrapper[4741]: I0929 20:53:10.736813 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:10 crc kubenswrapper[4741]: I0929 20:53:10.743997 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"cf1826cd-d5f3-4ed1-ad4e-9646edde3a65","Type":"ContainerStarted","Data":"3a172f338765283be38df455b22693c11436bc5adc74ead31836fdb1d37f64ae"} Sep 29 20:53:10 crc kubenswrapper[4741]: I0929 20:53:10.755600 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" podStartSLOduration=3.755586133 podStartE2EDuration="3.755586133s" podCreationTimestamp="2025-09-29 20:53:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:53:10.753497997 +0000 UTC m=+6232.401287319" watchObservedRunningTime="2025-09-29 20:53:10.755586133 +0000 UTC m=+6232.403375475" Sep 29 20:53:10 crc kubenswrapper[4741]: I0929 20:53:10.772957 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.104475799 podStartE2EDuration="3.77293904s" podCreationTimestamp="2025-09-29 20:53:07 +0000 UTC" firstStartedPulling="2025-09-29 20:53:08.319702975 +0000 UTC m=+6229.967492297" lastFinishedPulling="2025-09-29 20:53:08.988166206 +0000 UTC m=+6230.635955538" observedRunningTime="2025-09-29 20:53:10.770141992 +0000 UTC m=+6232.417931344" watchObservedRunningTime="2025-09-29 20:53:10.77293904 +0000 UTC m=+6232.420728372" Sep 29 20:53:11 crc kubenswrapper[4741]: I0929 20:53:11.085963 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:53:11 crc kubenswrapper[4741]: E0929 20:53:11.086313 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:53:14 crc kubenswrapper[4741]: I0929 20:53:14.792822 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a28a5999-6348-4d83-9605-16292704fb39","Type":"ContainerStarted","Data":"a0a0a75d640e61cb082acdc21047a8030e5c44be45e02b1e2da60eccffc77f91"} Sep 29 20:53:15 crc kubenswrapper[4741]: I0929 20:53:15.026867 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-d9twl"] Sep 29 20:53:15 crc kubenswrapper[4741]: I0929 20:53:15.036548 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-d9twl"] Sep 29 20:53:15 crc kubenswrapper[4741]: I0929 20:53:15.097262 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0dd2e44-2577-4579-911e-6a75c3063da0" path="/var/lib/kubelet/pods/a0dd2e44-2577-4579-911e-6a75c3063da0/volumes" Sep 29 20:53:15 crc kubenswrapper[4741]: I0929 20:53:15.805089 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a28a5999-6348-4d83-9605-16292704fb39","Type":"ContainerStarted","Data":"f5e1882bac451816cf6a7b0323406bad64cab20decd934854865b913f824f812"} Sep 29 20:53:15 crc kubenswrapper[4741]: I0929 20:53:15.827293 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.276162216 podStartE2EDuration="8.827269735s" podCreationTimestamp="2025-09-29 20:53:07 +0000 UTC" firstStartedPulling="2025-09-29 20:53:08.465434954 +0000 UTC m=+6230.113224286" lastFinishedPulling="2025-09-29 20:53:14.016542483 +0000 UTC m=+6235.664331805" observedRunningTime="2025-09-29 20:53:15.823560698 +0000 UTC m=+6237.471350040" watchObservedRunningTime="2025-09-29 20:53:15.827269735 +0000 UTC m=+6237.475059067" Sep 29 20:53:17 crc kubenswrapper[4741]: I0929 20:53:17.464129 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Sep 29 20:53:17 crc kubenswrapper[4741]: I0929 20:53:17.513956 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Sep 29 20:53:17 crc kubenswrapper[4741]: I0929 20:53:17.641351 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:53:17 crc kubenswrapper[4741]: I0929 20:53:17.716539 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866bddffb7-6rh5n"] Sep 29 20:53:17 crc kubenswrapper[4741]: I0929 20:53:17.716832 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" podUID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerName="dnsmasq-dns" containerID="cri-o://a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a" gracePeriod=10 Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.257997 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.396832 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-nb\") pod \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.396912 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-sb\") pod \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.397093 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-config\") pod \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.397179 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-dns-svc\") pod \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.397262 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxk7c\" (UniqueName: \"kubernetes.io/projected/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-kube-api-access-xxk7c\") pod \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\" (UID: \"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3\") " Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.417708 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-kube-api-access-xxk7c" (OuterVolumeSpecName: "kube-api-access-xxk7c") pod "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" (UID: "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3"). InnerVolumeSpecName "kube-api-access-xxk7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.466158 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" (UID: "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.472102 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" (UID: "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.476864 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-config" (OuterVolumeSpecName: "config") pod "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" (UID: "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.484913 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" (UID: "6d18f2ab-304c-4215-821a-5a1f8c8e0cb3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.500419 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.500445 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxk7c\" (UniqueName: \"kubernetes.io/projected/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-kube-api-access-xxk7c\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.500456 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.500465 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.500473 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.836131 4741 generic.go:334] "Generic (PLEG): container finished" podID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerID="a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a" exitCode=0 Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.836192 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" event={"ID":"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3","Type":"ContainerDied","Data":"a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a"} Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.836265 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" event={"ID":"6d18f2ab-304c-4215-821a-5a1f8c8e0cb3","Type":"ContainerDied","Data":"4bb816675ff295e7de37a28814d4c06fe477f2b6727c1f3e4ff1bdddada65124"} Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.836287 4741 scope.go:117] "RemoveContainer" containerID="a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.836702 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866bddffb7-6rh5n" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.912984 4741 scope.go:117] "RemoveContainer" containerID="6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.918437 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866bddffb7-6rh5n"] Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.932773 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-866bddffb7-6rh5n"] Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.934685 4741 scope.go:117] "RemoveContainer" containerID="a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a" Sep 29 20:53:18 crc kubenswrapper[4741]: E0929 20:53:18.935594 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a\": container with ID starting with a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a not found: ID does not exist" containerID="a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.935870 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a"} err="failed to get container status \"a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a\": rpc error: code = NotFound desc = could not find container \"a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a\": container with ID starting with a1267874fc5cec11e7486bf84c24397524512bfb87fd24880cbc29f73021c30a not found: ID does not exist" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.935896 4741 scope.go:117] "RemoveContainer" containerID="6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe" Sep 29 20:53:18 crc kubenswrapper[4741]: E0929 20:53:18.936840 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe\": container with ID starting with 6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe not found: ID does not exist" containerID="6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe" Sep 29 20:53:18 crc kubenswrapper[4741]: I0929 20:53:18.936966 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe"} err="failed to get container status \"6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe\": rpc error: code = NotFound desc = could not find container \"6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe\": container with ID starting with 6f68cb033501e48f785aa7ec16a59466a6853960fa859130da88d9ba7bc623fe not found: ID does not exist" Sep 29 20:53:19 crc kubenswrapper[4741]: I0929 20:53:19.100193 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" path="/var/lib/kubelet/pods/6d18f2ab-304c-4215-821a-5a1f8c8e0cb3/volumes" Sep 29 20:53:20 crc kubenswrapper[4741]: I0929 20:53:20.878378 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:53:20 crc kubenswrapper[4741]: I0929 20:53:20.878983 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-central-agent" containerID="cri-o://da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616" gracePeriod=30 Sep 29 20:53:20 crc kubenswrapper[4741]: I0929 20:53:20.879049 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-notification-agent" containerID="cri-o://b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218" gracePeriod=30 Sep 29 20:53:20 crc kubenswrapper[4741]: I0929 20:53:20.879050 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="sg-core" containerID="cri-o://e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e" gracePeriod=30 Sep 29 20:53:20 crc kubenswrapper[4741]: I0929 20:53:20.879050 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="proxy-httpd" containerID="cri-o://31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1" gracePeriod=30 Sep 29 20:53:21 crc kubenswrapper[4741]: I0929 20:53:21.903128 4741 generic.go:334] "Generic (PLEG): container finished" podID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerID="31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1" exitCode=0 Sep 29 20:53:21 crc kubenswrapper[4741]: I0929 20:53:21.903454 4741 generic.go:334] "Generic (PLEG): container finished" podID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerID="e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e" exitCode=2 Sep 29 20:53:21 crc kubenswrapper[4741]: I0929 20:53:21.903463 4741 generic.go:334] "Generic (PLEG): container finished" podID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerID="da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616" exitCode=0 Sep 29 20:53:21 crc kubenswrapper[4741]: I0929 20:53:21.903179 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerDied","Data":"31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1"} Sep 29 20:53:21 crc kubenswrapper[4741]: I0929 20:53:21.903499 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerDied","Data":"e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e"} Sep 29 20:53:21 crc kubenswrapper[4741]: I0929 20:53:21.903513 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerDied","Data":"da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616"} Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.452192 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.578104 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-scripts\") pod \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.578219 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-run-httpd\") pod \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.578295 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-combined-ca-bundle\") pod \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.578348 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-log-httpd\") pod \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.578410 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-sg-core-conf-yaml\") pod \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.578427 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-config-data\") pod \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.578556 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmt2p\" (UniqueName: \"kubernetes.io/projected/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-kube-api-access-pmt2p\") pod \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\" (UID: \"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d\") " Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.579246 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" (UID: "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.579421 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" (UID: "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.583361 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-kube-api-access-pmt2p" (OuterVolumeSpecName: "kube-api-access-pmt2p") pod "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" (UID: "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d"). InnerVolumeSpecName "kube-api-access-pmt2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.584261 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-scripts" (OuterVolumeSpecName: "scripts") pod "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" (UID: "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.613205 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" (UID: "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.660045 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" (UID: "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.682580 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmt2p\" (UniqueName: \"kubernetes.io/projected/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-kube-api-access-pmt2p\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.682608 4741 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.682617 4741 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.682627 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.682635 4741 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.682644 4741 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.715262 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-config-data" (OuterVolumeSpecName: "config-data") pod "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" (UID: "0ee89ee6-ef1a-40ee-9df7-0068cb434e4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.784065 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.917001 4741 generic.go:334] "Generic (PLEG): container finished" podID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerID="b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218" exitCode=0 Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.917352 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerDied","Data":"b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218"} Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.917384 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ee89ee6-ef1a-40ee-9df7-0068cb434e4d","Type":"ContainerDied","Data":"957341d5211bdaa10972588063c162e0aecf8719bb4f78d53f518d596628683b"} Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.917422 4741 scope.go:117] "RemoveContainer" containerID="31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.917595 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:53:22 crc kubenswrapper[4741]: I0929 20:53:22.974583 4741 scope.go:117] "RemoveContainer" containerID="e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.029912 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.042755 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054089 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.054628 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerName="init" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054649 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerName="init" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.054675 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="sg-core" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054684 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="sg-core" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.054702 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-central-agent" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054708 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-central-agent" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.054729 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-notification-agent" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054736 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-notification-agent" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.054747 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerName="dnsmasq-dns" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054753 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerName="dnsmasq-dns" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.054764 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="proxy-httpd" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054769 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="proxy-httpd" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054981 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d18f2ab-304c-4215-821a-5a1f8c8e0cb3" containerName="dnsmasq-dns" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.054996 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-central-agent" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.055012 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="sg-core" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.055029 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="ceilometer-notification-agent" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.055041 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" containerName="proxy-httpd" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.057155 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.061614 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.067740 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.069050 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.092149 4741 scope.go:117] "RemoveContainer" containerID="b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.102531 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ee89ee6-ef1a-40ee-9df7-0068cb434e4d" path="/var/lib/kubelet/pods/0ee89ee6-ef1a-40ee-9df7-0068cb434e4d/volumes" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.113248 4741 scope.go:117] "RemoveContainer" containerID="da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.135215 4741 scope.go:117] "RemoveContainer" containerID="31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.135689 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1\": container with ID starting with 31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1 not found: ID does not exist" containerID="31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.135718 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1"} err="failed to get container status \"31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1\": rpc error: code = NotFound desc = could not find container \"31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1\": container with ID starting with 31c6e65e89cc4e7625e90afb67552fb42df619335a7bed1f371c52c928b5eec1 not found: ID does not exist" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.135741 4741 scope.go:117] "RemoveContainer" containerID="e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.136303 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e\": container with ID starting with e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e not found: ID does not exist" containerID="e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.136357 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e"} err="failed to get container status \"e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e\": rpc error: code = NotFound desc = could not find container \"e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e\": container with ID starting with e1fe851ab4a2bc12f0a8fef3fee2d78f4d0286ddbfd9007143c59ae4e760fd5e not found: ID does not exist" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.136423 4741 scope.go:117] "RemoveContainer" containerID="b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.136757 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218\": container with ID starting with b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218 not found: ID does not exist" containerID="b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.136781 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218"} err="failed to get container status \"b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218\": rpc error: code = NotFound desc = could not find container \"b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218\": container with ID starting with b84b710ae3174025722bdb7655ad31e166c1439e86b1e2dafb91014e0a98b218 not found: ID does not exist" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.136795 4741 scope.go:117] "RemoveContainer" containerID="da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.137319 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616\": container with ID starting with da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616 not found: ID does not exist" containerID="da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.137355 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616"} err="failed to get container status \"da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616\": rpc error: code = NotFound desc = could not find container \"da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616\": container with ID starting with da866cfd6a0a71905fa50df020f445a125323faa3371b44b03c0cfdb3c7f7616 not found: ID does not exist" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.192590 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5194444-589f-4132-8696-ccac62962087-run-httpd\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.192657 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-scripts\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.192695 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-config-data\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.192731 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.192866 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5194444-589f-4132-8696-ccac62962087-log-httpd\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.192918 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.193022 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6krt\" (UniqueName: \"kubernetes.io/projected/d5194444-589f-4132-8696-ccac62962087-kube-api-access-x6krt\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: E0929 20:53:23.241403 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ee89ee6_ef1a_40ee_9df7_0068cb434e4d.slice/crio-957341d5211bdaa10972588063c162e0aecf8719bb4f78d53f518d596628683b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ee89ee6_ef1a_40ee_9df7_0068cb434e4d.slice\": RecentStats: unable to find data in memory cache]" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.296369 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6krt\" (UniqueName: \"kubernetes.io/projected/d5194444-589f-4132-8696-ccac62962087-kube-api-access-x6krt\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.296471 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5194444-589f-4132-8696-ccac62962087-run-httpd\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.296521 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-scripts\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.296563 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-config-data\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.296611 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.296649 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5194444-589f-4132-8696-ccac62962087-log-httpd\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.296723 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.297826 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5194444-589f-4132-8696-ccac62962087-log-httpd\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.297891 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5194444-589f-4132-8696-ccac62962087-run-httpd\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.301695 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.301969 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.302837 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-config-data\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.313463 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5194444-589f-4132-8696-ccac62962087-scripts\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.314052 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6krt\" (UniqueName: \"kubernetes.io/projected/d5194444-589f-4132-8696-ccac62962087-kube-api-access-x6krt\") pod \"ceilometer-0\" (UID: \"d5194444-589f-4132-8696-ccac62962087\") " pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.392236 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.866938 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 20:53:23 crc kubenswrapper[4741]: W0929 20:53:23.872649 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5194444_589f_4132_8696_ccac62962087.slice/crio-dc5466c73780eeda0e2be5c6f6d224b7e12397a41a3457477f91f7f248ce6199 WatchSource:0}: Error finding container dc5466c73780eeda0e2be5c6f6d224b7e12397a41a3457477f91f7f248ce6199: Status 404 returned error can't find the container with id dc5466c73780eeda0e2be5c6f6d224b7e12397a41a3457477f91f7f248ce6199 Sep 29 20:53:23 crc kubenswrapper[4741]: I0929 20:53:23.930211 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5194444-589f-4132-8696-ccac62962087","Type":"ContainerStarted","Data":"dc5466c73780eeda0e2be5c6f6d224b7e12397a41a3457477f91f7f248ce6199"} Sep 29 20:53:24 crc kubenswrapper[4741]: I0929 20:53:24.086616 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:53:24 crc kubenswrapper[4741]: E0929 20:53:24.086870 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 20:53:24 crc kubenswrapper[4741]: I0929 20:53:24.946572 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5194444-589f-4132-8696-ccac62962087","Type":"ContainerStarted","Data":"c06b1ae202891f1e3de749321800eeb914efcf619b1b7abfc0b6eb00d8e4311d"} Sep 29 20:53:25 crc kubenswrapper[4741]: I0929 20:53:25.961729 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5194444-589f-4132-8696-ccac62962087","Type":"ContainerStarted","Data":"c1a690197742debd7350969e423970b7675206ad9adb749e94214e3e9d6b1fe8"} Sep 29 20:53:26 crc kubenswrapper[4741]: I0929 20:53:26.972688 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5194444-589f-4132-8696-ccac62962087","Type":"ContainerStarted","Data":"6b790bf80d1857a0ca1242930b097921a182db41121262ae439916e6407830bd"} Sep 29 20:53:28 crc kubenswrapper[4741]: I0929 20:53:28.984733 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Sep 29 20:53:28 crc kubenswrapper[4741]: I0929 20:53:28.998924 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5194444-589f-4132-8696-ccac62962087","Type":"ContainerStarted","Data":"e0f7428007b05b02f1a2acbb4c8ff844d304e633c00eae75401242519be93f96"} Sep 29 20:53:28 crc kubenswrapper[4741]: I0929 20:53:28.999425 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.052829 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.976084078 podStartE2EDuration="7.052813078s" podCreationTimestamp="2025-09-29 20:53:22 +0000 UTC" firstStartedPulling="2025-09-29 20:53:23.87516346 +0000 UTC m=+6245.522952792" lastFinishedPulling="2025-09-29 20:53:27.95189246 +0000 UTC m=+6249.599681792" observedRunningTime="2025-09-29 20:53:29.041185273 +0000 UTC m=+6250.688974605" watchObservedRunningTime="2025-09-29 20:53:29.052813078 +0000 UTC m=+6250.700602410" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.143792 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.410167 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.717985 4741 scope.go:117] "RemoveContainer" containerID="fbf8f9b90c9606b691caf816febd66a5f65f40a818d40e4b04d6ba134fc57f43" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.740369 4741 scope.go:117] "RemoveContainer" containerID="0ec01ac2e2528dd738c8abcd16735488f6b79156508f049af4d769d31f1396b7" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.814094 4741 scope.go:117] "RemoveContainer" containerID="4db70c09719aa843b9490e6f59cba797fa57f71c212f4752942fcd5447a1f1e9" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.847467 4741 scope.go:117] "RemoveContainer" containerID="27c279c151855c3f97e1ff2e9ad21c728949a8336c84380bee8b4e58ef7fad93" Sep 29 20:53:29 crc kubenswrapper[4741]: I0929 20:53:29.910086 4741 scope.go:117] "RemoveContainer" containerID="841798414a0df3202bdd06741118227e51058c7cf49d891fd766cdd9ee9464f1" Sep 29 20:53:37 crc kubenswrapper[4741]: I0929 20:53:37.090155 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:53:38 crc kubenswrapper[4741]: I0929 20:53:38.099162 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"9532bce3d47878ef326b645cc834aaa5500a0112cd2409697bc4431be4e30531"} Sep 29 20:53:53 crc kubenswrapper[4741]: I0929 20:53:53.400446 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.423966 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7789547b9c-76cfr"] Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.427522 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.431886 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.443613 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7789547b9c-76cfr"] Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.597739 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-config\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.598111 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stfrn\" (UniqueName: \"kubernetes.io/projected/0bf19132-9205-44a9-91e5-c5433351f6eb-kube-api-access-stfrn\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.598223 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-sb\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.598262 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-nb\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.598332 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-openstack-cell1\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.598542 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-dns-svc\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.699885 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-dns-svc\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.700033 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-config\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.700056 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stfrn\" (UniqueName: \"kubernetes.io/projected/0bf19132-9205-44a9-91e5-c5433351f6eb-kube-api-access-stfrn\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.700083 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-sb\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.700106 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-nb\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.700162 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-openstack-cell1\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.701122 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-sb\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.701127 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-dns-svc\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.701184 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-openstack-cell1\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.701210 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-nb\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.701719 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-config\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.728155 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stfrn\" (UniqueName: \"kubernetes.io/projected/0bf19132-9205-44a9-91e5-c5433351f6eb-kube-api-access-stfrn\") pod \"dnsmasq-dns-7789547b9c-76cfr\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:18 crc kubenswrapper[4741]: I0929 20:54:18.758880 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:19 crc kubenswrapper[4741]: W0929 20:54:19.284338 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bf19132_9205_44a9_91e5_c5433351f6eb.slice/crio-ccbf1a20ced99a3c73af83e9fc813a08dcc56c656aac9d55b2e9506c3f20d72a WatchSource:0}: Error finding container ccbf1a20ced99a3c73af83e9fc813a08dcc56c656aac9d55b2e9506c3f20d72a: Status 404 returned error can't find the container with id ccbf1a20ced99a3c73af83e9fc813a08dcc56c656aac9d55b2e9506c3f20d72a Sep 29 20:54:19 crc kubenswrapper[4741]: I0929 20:54:19.284556 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7789547b9c-76cfr"] Sep 29 20:54:19 crc kubenswrapper[4741]: I0929 20:54:19.600217 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" event={"ID":"0bf19132-9205-44a9-91e5-c5433351f6eb","Type":"ContainerStarted","Data":"647d5e5a59c9b4b387af8969c28ddf5211f8a0a24b30831fcd87fadbfe9d7004"} Sep 29 20:54:19 crc kubenswrapper[4741]: I0929 20:54:19.600524 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" event={"ID":"0bf19132-9205-44a9-91e5-c5433351f6eb","Type":"ContainerStarted","Data":"ccbf1a20ced99a3c73af83e9fc813a08dcc56c656aac9d55b2e9506c3f20d72a"} Sep 29 20:54:20 crc kubenswrapper[4741]: I0929 20:54:20.611026 4741 generic.go:334] "Generic (PLEG): container finished" podID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerID="647d5e5a59c9b4b387af8969c28ddf5211f8a0a24b30831fcd87fadbfe9d7004" exitCode=0 Sep 29 20:54:20 crc kubenswrapper[4741]: I0929 20:54:20.611278 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" event={"ID":"0bf19132-9205-44a9-91e5-c5433351f6eb","Type":"ContainerDied","Data":"647d5e5a59c9b4b387af8969c28ddf5211f8a0a24b30831fcd87fadbfe9d7004"} Sep 29 20:54:21 crc kubenswrapper[4741]: I0929 20:54:21.623977 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" event={"ID":"0bf19132-9205-44a9-91e5-c5433351f6eb","Type":"ContainerStarted","Data":"b86ce3936d8731243ff7b70216082259a6dbdf899cc96c3e442cd4beb0bbb62b"} Sep 29 20:54:21 crc kubenswrapper[4741]: I0929 20:54:21.624494 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:21 crc kubenswrapper[4741]: I0929 20:54:21.656529 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" podStartSLOduration=3.6565036060000002 podStartE2EDuration="3.656503606s" podCreationTimestamp="2025-09-29 20:54:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:54:21.650041443 +0000 UTC m=+6303.297830785" watchObservedRunningTime="2025-09-29 20:54:21.656503606 +0000 UTC m=+6303.304292948" Sep 29 20:54:28 crc kubenswrapper[4741]: I0929 20:54:28.760568 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:28 crc kubenswrapper[4741]: I0929 20:54:28.831822 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bf58c57-x7s92"] Sep 29 20:54:28 crc kubenswrapper[4741]: I0929 20:54:28.832158 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" podUID="639e8429-8817-42f6-bdea-1a96545e6640" containerName="dnsmasq-dns" containerID="cri-o://608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259" gracePeriod=10 Sep 29 20:54:28 crc kubenswrapper[4741]: I0929 20:54:28.952443 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5995b898cf-xwc7d"] Sep 29 20:54:28 crc kubenswrapper[4741]: I0929 20:54:28.954537 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:28 crc kubenswrapper[4741]: I0929 20:54:28.960420 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5995b898cf-xwc7d"] Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.085090 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-ovsdbserver-nb\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.085147 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-openstack-cell1\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.085213 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-config\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.085334 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpbw9\" (UniqueName: \"kubernetes.io/projected/cb3a3160-8aa9-49f5-9479-208262519896-kube-api-access-kpbw9\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.085360 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-ovsdbserver-sb\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.085384 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-dns-svc\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.187574 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-ovsdbserver-nb\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.187909 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-openstack-cell1\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.187986 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-config\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.188124 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpbw9\" (UniqueName: \"kubernetes.io/projected/cb3a3160-8aa9-49f5-9479-208262519896-kube-api-access-kpbw9\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.188147 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-ovsdbserver-sb\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.188178 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-dns-svc\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.190191 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-dns-svc\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.190837 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-ovsdbserver-nb\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.191472 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-openstack-cell1\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.192142 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-config\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.193187 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3a3160-8aa9-49f5-9479-208262519896-ovsdbserver-sb\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.223503 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpbw9\" (UniqueName: \"kubernetes.io/projected/cb3a3160-8aa9-49f5-9479-208262519896-kube-api-access-kpbw9\") pod \"dnsmasq-dns-5995b898cf-xwc7d\" (UID: \"cb3a3160-8aa9-49f5-9479-208262519896\") " pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.301144 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.515622 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.707963 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jdms\" (UniqueName: \"kubernetes.io/projected/639e8429-8817-42f6-bdea-1a96545e6640-kube-api-access-8jdms\") pod \"639e8429-8817-42f6-bdea-1a96545e6640\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.708119 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-sb\") pod \"639e8429-8817-42f6-bdea-1a96545e6640\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.708267 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-config\") pod \"639e8429-8817-42f6-bdea-1a96545e6640\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.708344 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-nb\") pod \"639e8429-8817-42f6-bdea-1a96545e6640\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.708979 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-dns-svc\") pod \"639e8429-8817-42f6-bdea-1a96545e6640\" (UID: \"639e8429-8817-42f6-bdea-1a96545e6640\") " Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.713580 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/639e8429-8817-42f6-bdea-1a96545e6640-kube-api-access-8jdms" (OuterVolumeSpecName: "kube-api-access-8jdms") pod "639e8429-8817-42f6-bdea-1a96545e6640" (UID: "639e8429-8817-42f6-bdea-1a96545e6640"). InnerVolumeSpecName "kube-api-access-8jdms". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.714517 4741 generic.go:334] "Generic (PLEG): container finished" podID="639e8429-8817-42f6-bdea-1a96545e6640" containerID="608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259" exitCode=0 Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.714560 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" event={"ID":"639e8429-8817-42f6-bdea-1a96545e6640","Type":"ContainerDied","Data":"608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259"} Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.714587 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" event={"ID":"639e8429-8817-42f6-bdea-1a96545e6640","Type":"ContainerDied","Data":"e711b0df64968ad06334c38425a9ae66674de6146349d4b09aebcccdf145e917"} Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.714603 4741 scope.go:117] "RemoveContainer" containerID="608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.714704 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bf58c57-x7s92" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.745641 4741 scope.go:117] "RemoveContainer" containerID="2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.790340 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "639e8429-8817-42f6-bdea-1a96545e6640" (UID: "639e8429-8817-42f6-bdea-1a96545e6640"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.790883 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "639e8429-8817-42f6-bdea-1a96545e6640" (UID: "639e8429-8817-42f6-bdea-1a96545e6640"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.797641 4741 scope.go:117] "RemoveContainer" containerID="608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259" Sep 29 20:54:29 crc kubenswrapper[4741]: E0929 20:54:29.798147 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259\": container with ID starting with 608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259 not found: ID does not exist" containerID="608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.798205 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259"} err="failed to get container status \"608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259\": rpc error: code = NotFound desc = could not find container \"608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259\": container with ID starting with 608f8c394fe08f014f5f348b92cef8f942516ef0450d2a0d6b96bf54bb680259 not found: ID does not exist" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.798251 4741 scope.go:117] "RemoveContainer" containerID="2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c" Sep 29 20:54:29 crc kubenswrapper[4741]: E0929 20:54:29.798744 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c\": container with ID starting with 2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c not found: ID does not exist" containerID="2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.798864 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c"} err="failed to get container status \"2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c\": rpc error: code = NotFound desc = could not find container \"2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c\": container with ID starting with 2100e40a63609cb7af4ef2fe82d507f4d0b44613baeddb527113d0a575f8a30c not found: ID does not exist" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.803790 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "639e8429-8817-42f6-bdea-1a96545e6640" (UID: "639e8429-8817-42f6-bdea-1a96545e6640"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.812217 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jdms\" (UniqueName: \"kubernetes.io/projected/639e8429-8817-42f6-bdea-1a96545e6640-kube-api-access-8jdms\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.812247 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.812256 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.812267 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.820027 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-config" (OuterVolumeSpecName: "config") pod "639e8429-8817-42f6-bdea-1a96545e6640" (UID: "639e8429-8817-42f6-bdea-1a96545e6640"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.896198 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5995b898cf-xwc7d"] Sep 29 20:54:29 crc kubenswrapper[4741]: I0929 20:54:29.913753 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/639e8429-8817-42f6-bdea-1a96545e6640-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:30 crc kubenswrapper[4741]: I0929 20:54:30.181474 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bf58c57-x7s92"] Sep 29 20:54:30 crc kubenswrapper[4741]: I0929 20:54:30.197462 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-797bf58c57-x7s92"] Sep 29 20:54:30 crc kubenswrapper[4741]: I0929 20:54:30.724198 4741 generic.go:334] "Generic (PLEG): container finished" podID="cb3a3160-8aa9-49f5-9479-208262519896" containerID="41a7a745f35e5218e2c97053ee713e856206800f5c87fa3e4c99da88872e88e4" exitCode=0 Sep 29 20:54:30 crc kubenswrapper[4741]: I0929 20:54:30.724244 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" event={"ID":"cb3a3160-8aa9-49f5-9479-208262519896","Type":"ContainerDied","Data":"41a7a745f35e5218e2c97053ee713e856206800f5c87fa3e4c99da88872e88e4"} Sep 29 20:54:30 crc kubenswrapper[4741]: I0929 20:54:30.724844 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" event={"ID":"cb3a3160-8aa9-49f5-9479-208262519896","Type":"ContainerStarted","Data":"a449f7258e253c75ae9aac6071603bedac7cac830447ddcdcfa93a4dcc89509f"} Sep 29 20:54:31 crc kubenswrapper[4741]: I0929 20:54:31.096956 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="639e8429-8817-42f6-bdea-1a96545e6640" path="/var/lib/kubelet/pods/639e8429-8817-42f6-bdea-1a96545e6640/volumes" Sep 29 20:54:31 crc kubenswrapper[4741]: I0929 20:54:31.734465 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" event={"ID":"cb3a3160-8aa9-49f5-9479-208262519896","Type":"ContainerStarted","Data":"f36d96606043b0cbd3c6ab78fc7c25552757f9344aa9d8a3bf34f6cf2f397ad4"} Sep 29 20:54:31 crc kubenswrapper[4741]: I0929 20:54:31.734626 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:31 crc kubenswrapper[4741]: I0929 20:54:31.756417 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" podStartSLOduration=3.756400361 podStartE2EDuration="3.756400361s" podCreationTimestamp="2025-09-29 20:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 20:54:31.749845424 +0000 UTC m=+6313.397634756" watchObservedRunningTime="2025-09-29 20:54:31.756400361 +0000 UTC m=+6313.404189693" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.751920 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85"] Sep 29 20:54:35 crc kubenswrapper[4741]: E0929 20:54:35.754085 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639e8429-8817-42f6-bdea-1a96545e6640" containerName="dnsmasq-dns" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.755110 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="639e8429-8817-42f6-bdea-1a96545e6640" containerName="dnsmasq-dns" Sep 29 20:54:35 crc kubenswrapper[4741]: E0929 20:54:35.755263 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639e8429-8817-42f6-bdea-1a96545e6640" containerName="init" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.755349 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="639e8429-8817-42f6-bdea-1a96545e6640" containerName="init" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.755743 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="639e8429-8817-42f6-bdea-1a96545e6640" containerName="dnsmasq-dns" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.756779 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.760972 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.761390 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.761741 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.761937 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.778709 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85"] Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.854019 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.854104 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h5bk\" (UniqueName: \"kubernetes.io/projected/3e541050-007b-407c-9a45-2eb70ade11d9-kube-api-access-8h5bk\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.854142 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.854433 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.854502 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.956016 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.956072 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.956203 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.956286 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h5bk\" (UniqueName: \"kubernetes.io/projected/3e541050-007b-407c-9a45-2eb70ade11d9-kube-api-access-8h5bk\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.956331 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.963204 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.963260 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.963979 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.964266 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:35 crc kubenswrapper[4741]: I0929 20:54:35.972346 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h5bk\" (UniqueName: \"kubernetes.io/projected/3e541050-007b-407c-9a45-2eb70ade11d9-kube-api-access-8h5bk\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:36 crc kubenswrapper[4741]: I0929 20:54:36.104563 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:54:36 crc kubenswrapper[4741]: I0929 20:54:36.663074 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85"] Sep 29 20:54:36 crc kubenswrapper[4741]: W0929 20:54:36.667936 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e541050_007b_407c_9a45_2eb70ade11d9.slice/crio-6e0a47785759dcdbc5777118aa73ad153da0efd8a0ecd9ea2e910f5ba853c9db WatchSource:0}: Error finding container 6e0a47785759dcdbc5777118aa73ad153da0efd8a0ecd9ea2e910f5ba853c9db: Status 404 returned error can't find the container with id 6e0a47785759dcdbc5777118aa73ad153da0efd8a0ecd9ea2e910f5ba853c9db Sep 29 20:54:36 crc kubenswrapper[4741]: I0929 20:54:36.671078 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:54:36 crc kubenswrapper[4741]: I0929 20:54:36.824364 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" event={"ID":"3e541050-007b-407c-9a45-2eb70ade11d9","Type":"ContainerStarted","Data":"6e0a47785759dcdbc5777118aa73ad153da0efd8a0ecd9ea2e910f5ba853c9db"} Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.303299 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5995b898cf-xwc7d" Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.363376 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7789547b9c-76cfr"] Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.363604 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" podUID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerName="dnsmasq-dns" containerID="cri-o://b86ce3936d8731243ff7b70216082259a6dbdf899cc96c3e442cd4beb0bbb62b" gracePeriod=10 Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.866853 4741 generic.go:334] "Generic (PLEG): container finished" podID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerID="b86ce3936d8731243ff7b70216082259a6dbdf899cc96c3e442cd4beb0bbb62b" exitCode=0 Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.866937 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" event={"ID":"0bf19132-9205-44a9-91e5-c5433351f6eb","Type":"ContainerDied","Data":"b86ce3936d8731243ff7b70216082259a6dbdf899cc96c3e442cd4beb0bbb62b"} Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.867164 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" event={"ID":"0bf19132-9205-44a9-91e5-c5433351f6eb","Type":"ContainerDied","Data":"ccbf1a20ced99a3c73af83e9fc813a08dcc56c656aac9d55b2e9506c3f20d72a"} Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.867181 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccbf1a20ced99a3c73af83e9fc813a08dcc56c656aac9d55b2e9506c3f20d72a" Sep 29 20:54:39 crc kubenswrapper[4741]: I0929 20:54:39.952910 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.052413 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-config\") pod \"0bf19132-9205-44a9-91e5-c5433351f6eb\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.052497 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stfrn\" (UniqueName: \"kubernetes.io/projected/0bf19132-9205-44a9-91e5-c5433351f6eb-kube-api-access-stfrn\") pod \"0bf19132-9205-44a9-91e5-c5433351f6eb\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.052605 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-dns-svc\") pod \"0bf19132-9205-44a9-91e5-c5433351f6eb\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.052694 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-openstack-cell1\") pod \"0bf19132-9205-44a9-91e5-c5433351f6eb\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.052784 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-nb\") pod \"0bf19132-9205-44a9-91e5-c5433351f6eb\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.052846 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-sb\") pod \"0bf19132-9205-44a9-91e5-c5433351f6eb\" (UID: \"0bf19132-9205-44a9-91e5-c5433351f6eb\") " Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.073723 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf19132-9205-44a9-91e5-c5433351f6eb-kube-api-access-stfrn" (OuterVolumeSpecName: "kube-api-access-stfrn") pod "0bf19132-9205-44a9-91e5-c5433351f6eb" (UID: "0bf19132-9205-44a9-91e5-c5433351f6eb"). InnerVolumeSpecName "kube-api-access-stfrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.112955 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-config" (OuterVolumeSpecName: "config") pod "0bf19132-9205-44a9-91e5-c5433351f6eb" (UID: "0bf19132-9205-44a9-91e5-c5433351f6eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.116010 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0bf19132-9205-44a9-91e5-c5433351f6eb" (UID: "0bf19132-9205-44a9-91e5-c5433351f6eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.127977 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0bf19132-9205-44a9-91e5-c5433351f6eb" (UID: "0bf19132-9205-44a9-91e5-c5433351f6eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.128598 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "0bf19132-9205-44a9-91e5-c5433351f6eb" (UID: "0bf19132-9205-44a9-91e5-c5433351f6eb"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.138858 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0bf19132-9205-44a9-91e5-c5433351f6eb" (UID: "0bf19132-9205-44a9-91e5-c5433351f6eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.156163 4741 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.156199 4741 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-openstack-cell1\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.156214 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.156225 4741 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.156235 4741 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf19132-9205-44a9-91e5-c5433351f6eb-config\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.156248 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stfrn\" (UniqueName: \"kubernetes.io/projected/0bf19132-9205-44a9-91e5-c5433351f6eb-kube-api-access-stfrn\") on node \"crc\" DevicePath \"\"" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.879219 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789547b9c-76cfr" Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.921567 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7789547b9c-76cfr"] Sep 29 20:54:40 crc kubenswrapper[4741]: I0929 20:54:40.932588 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7789547b9c-76cfr"] Sep 29 20:54:41 crc kubenswrapper[4741]: I0929 20:54:41.097239 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf19132-9205-44a9-91e5-c5433351f6eb" path="/var/lib/kubelet/pods/0bf19132-9205-44a9-91e5-c5433351f6eb/volumes" Sep 29 20:54:46 crc kubenswrapper[4741]: I0929 20:54:46.935589 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" event={"ID":"3e541050-007b-407c-9a45-2eb70ade11d9","Type":"ContainerStarted","Data":"92b6de23b8dfb58c2b04ea32b5e072df81b90229062dd27298dfa43a94e35527"} Sep 29 20:54:46 crc kubenswrapper[4741]: I0929 20:54:46.966009 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" podStartSLOduration=2.405282939 podStartE2EDuration="11.965991553s" podCreationTimestamp="2025-09-29 20:54:35 +0000 UTC" firstStartedPulling="2025-09-29 20:54:36.670772439 +0000 UTC m=+6318.318561781" lastFinishedPulling="2025-09-29 20:54:46.231481063 +0000 UTC m=+6327.879270395" observedRunningTime="2025-09-29 20:54:46.959522279 +0000 UTC m=+6328.607311611" watchObservedRunningTime="2025-09-29 20:54:46.965991553 +0000 UTC m=+6328.613780885" Sep 29 20:55:02 crc kubenswrapper[4741]: I0929 20:55:02.095290 4741 generic.go:334] "Generic (PLEG): container finished" podID="3e541050-007b-407c-9a45-2eb70ade11d9" containerID="92b6de23b8dfb58c2b04ea32b5e072df81b90229062dd27298dfa43a94e35527" exitCode=0 Sep 29 20:55:02 crc kubenswrapper[4741]: I0929 20:55:02.095344 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" event={"ID":"3e541050-007b-407c-9a45-2eb70ade11d9","Type":"ContainerDied","Data":"92b6de23b8dfb58c2b04ea32b5e072df81b90229062dd27298dfa43a94e35527"} Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.660277 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.687444 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-inventory\") pod \"3e541050-007b-407c-9a45-2eb70ade11d9\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.687515 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h5bk\" (UniqueName: \"kubernetes.io/projected/3e541050-007b-407c-9a45-2eb70ade11d9-kube-api-access-8h5bk\") pod \"3e541050-007b-407c-9a45-2eb70ade11d9\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.687559 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ssh-key\") pod \"3e541050-007b-407c-9a45-2eb70ade11d9\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.688534 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-pre-adoption-validation-combined-ca-bundle\") pod \"3e541050-007b-407c-9a45-2eb70ade11d9\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.688626 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ceph\") pod \"3e541050-007b-407c-9a45-2eb70ade11d9\" (UID: \"3e541050-007b-407c-9a45-2eb70ade11d9\") " Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.693496 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e541050-007b-407c-9a45-2eb70ade11d9-kube-api-access-8h5bk" (OuterVolumeSpecName: "kube-api-access-8h5bk") pod "3e541050-007b-407c-9a45-2eb70ade11d9" (UID: "3e541050-007b-407c-9a45-2eb70ade11d9"). InnerVolumeSpecName "kube-api-access-8h5bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.694138 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "3e541050-007b-407c-9a45-2eb70ade11d9" (UID: "3e541050-007b-407c-9a45-2eb70ade11d9"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.694400 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ceph" (OuterVolumeSpecName: "ceph") pod "3e541050-007b-407c-9a45-2eb70ade11d9" (UID: "3e541050-007b-407c-9a45-2eb70ade11d9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.723561 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3e541050-007b-407c-9a45-2eb70ade11d9" (UID: "3e541050-007b-407c-9a45-2eb70ade11d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.725399 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-inventory" (OuterVolumeSpecName: "inventory") pod "3e541050-007b-407c-9a45-2eb70ade11d9" (UID: "3e541050-007b-407c-9a45-2eb70ade11d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.791405 4741 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.791441 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.791454 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.791467 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h5bk\" (UniqueName: \"kubernetes.io/projected/3e541050-007b-407c-9a45-2eb70ade11d9-kube-api-access-8h5bk\") on node \"crc\" DevicePath \"\"" Sep 29 20:55:03 crc kubenswrapper[4741]: I0929 20:55:03.791500 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e541050-007b-407c-9a45-2eb70ade11d9-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 20:55:04 crc kubenswrapper[4741]: I0929 20:55:04.120094 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" event={"ID":"3e541050-007b-407c-9a45-2eb70ade11d9","Type":"ContainerDied","Data":"6e0a47785759dcdbc5777118aa73ad153da0efd8a0ecd9ea2e910f5ba853c9db"} Sep 29 20:55:04 crc kubenswrapper[4741]: I0929 20:55:04.120504 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e0a47785759dcdbc5777118aa73ad153da0efd8a0ecd9ea2e910f5ba853c9db" Sep 29 20:55:04 crc kubenswrapper[4741]: I0929 20:55:04.120602 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.983978 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5"] Sep 29 20:55:13 crc kubenswrapper[4741]: E0929 20:55:13.984999 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerName="dnsmasq-dns" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.985018 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerName="dnsmasq-dns" Sep 29 20:55:13 crc kubenswrapper[4741]: E0929 20:55:13.985078 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e541050-007b-407c-9a45-2eb70ade11d9" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.985092 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e541050-007b-407c-9a45-2eb70ade11d9" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Sep 29 20:55:13 crc kubenswrapper[4741]: E0929 20:55:13.985131 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerName="init" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.985141 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerName="init" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.985536 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf19132-9205-44a9-91e5-c5433351f6eb" containerName="dnsmasq-dns" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.985577 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e541050-007b-407c-9a45-2eb70ade11d9" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.988635 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.998478 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.998601 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.998478 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 20:55:13 crc kubenswrapper[4741]: I0929 20:55:13.999065 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.013603 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5"] Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.102864 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.102948 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksbhb\" (UniqueName: \"kubernetes.io/projected/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-kube-api-access-ksbhb\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.103163 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.103484 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.103675 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.205894 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.205937 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.205967 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksbhb\" (UniqueName: \"kubernetes.io/projected/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-kube-api-access-ksbhb\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.206006 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.206112 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.216373 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.216740 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.217354 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.217443 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.230563 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksbhb\" (UniqueName: \"kubernetes.io/projected/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-kube-api-access-ksbhb\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.317092 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 20:55:14 crc kubenswrapper[4741]: I0929 20:55:14.836188 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5"] Sep 29 20:55:14 crc kubenswrapper[4741]: W0929 20:55:14.841881 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3d3a768_7bc7_4c8c_ade1_f8ddcabd8261.slice/crio-4dea27e72a0012f7bd84a84bab136fe7f7995271186636c84e0126872691bf7e WatchSource:0}: Error finding container 4dea27e72a0012f7bd84a84bab136fe7f7995271186636c84e0126872691bf7e: Status 404 returned error can't find the container with id 4dea27e72a0012f7bd84a84bab136fe7f7995271186636c84e0126872691bf7e Sep 29 20:55:15 crc kubenswrapper[4741]: I0929 20:55:15.257124 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" event={"ID":"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261","Type":"ContainerStarted","Data":"4dea27e72a0012f7bd84a84bab136fe7f7995271186636c84e0126872691bf7e"} Sep 29 20:55:16 crc kubenswrapper[4741]: I0929 20:55:16.280632 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" event={"ID":"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261","Type":"ContainerStarted","Data":"5becc5f6051a5ac17b6b6184af7adee17f732b30c6122a71c44314a685589375"} Sep 29 20:55:16 crc kubenswrapper[4741]: I0929 20:55:16.341042 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" podStartSLOduration=2.915710866 podStartE2EDuration="3.341021289s" podCreationTimestamp="2025-09-29 20:55:13 +0000 UTC" firstStartedPulling="2025-09-29 20:55:14.845000959 +0000 UTC m=+6356.492790281" lastFinishedPulling="2025-09-29 20:55:15.270311372 +0000 UTC m=+6356.918100704" observedRunningTime="2025-09-29 20:55:16.312549204 +0000 UTC m=+6357.960338586" watchObservedRunningTime="2025-09-29 20:55:16.341021289 +0000 UTC m=+6357.988810641" Sep 29 20:56:01 crc kubenswrapper[4741]: I0929 20:56:01.042580 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-fsrvf"] Sep 29 20:56:01 crc kubenswrapper[4741]: I0929 20:56:01.052099 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-fsrvf"] Sep 29 20:56:01 crc kubenswrapper[4741]: I0929 20:56:01.101781 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deda0fe8-0c19-4799-a2de-45c6a10014cb" path="/var/lib/kubelet/pods/deda0fe8-0c19-4799-a2de-45c6a10014cb/volumes" Sep 29 20:56:01 crc kubenswrapper[4741]: I0929 20:56:01.738651 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:56:01 crc kubenswrapper[4741]: I0929 20:56:01.738743 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:56:13 crc kubenswrapper[4741]: I0929 20:56:13.045232 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-9899-account-create-5vxm7"] Sep 29 20:56:13 crc kubenswrapper[4741]: I0929 20:56:13.057721 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-9899-account-create-5vxm7"] Sep 29 20:56:13 crc kubenswrapper[4741]: I0929 20:56:13.102973 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b2fee03-161c-4a66-8abd-0d793e229153" path="/var/lib/kubelet/pods/3b2fee03-161c-4a66-8abd-0d793e229153/volumes" Sep 29 20:56:19 crc kubenswrapper[4741]: I0929 20:56:19.036101 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-rskwp"] Sep 29 20:56:19 crc kubenswrapper[4741]: I0929 20:56:19.047708 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-rskwp"] Sep 29 20:56:19 crc kubenswrapper[4741]: I0929 20:56:19.109004 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2e774cc-7135-4580-9f0d-82ead3f51c0b" path="/var/lib/kubelet/pods/e2e774cc-7135-4580-9f0d-82ead3f51c0b/volumes" Sep 29 20:56:30 crc kubenswrapper[4741]: I0929 20:56:30.227120 4741 scope.go:117] "RemoveContainer" containerID="12e42207162c5b9dae00c1f4d62d733c8d650ab7de27d525d7419ffb9c7a2859" Sep 29 20:56:30 crc kubenswrapper[4741]: I0929 20:56:30.274059 4741 scope.go:117] "RemoveContainer" containerID="a95b5600cb9cc4cf69f11a889eff3455aaa0b97c8a992ebec97d924cae316651" Sep 29 20:56:30 crc kubenswrapper[4741]: I0929 20:56:30.322180 4741 scope.go:117] "RemoveContainer" containerID="700a401ff67817edf51f2c71f55522cf1c31134d6d9bce8963cebd0340e9955a" Sep 29 20:56:31 crc kubenswrapper[4741]: I0929 20:56:31.047592 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-f535-account-create-drz2b"] Sep 29 20:56:31 crc kubenswrapper[4741]: I0929 20:56:31.059210 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-f535-account-create-drz2b"] Sep 29 20:56:31 crc kubenswrapper[4741]: I0929 20:56:31.101536 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb715a13-36f6-4750-baa3-1f51eeb3856d" path="/var/lib/kubelet/pods/cb715a13-36f6-4750-baa3-1f51eeb3856d/volumes" Sep 29 20:56:31 crc kubenswrapper[4741]: I0929 20:56:31.739429 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:56:31 crc kubenswrapper[4741]: I0929 20:56:31.739819 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:57:01 crc kubenswrapper[4741]: I0929 20:57:01.738771 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:57:01 crc kubenswrapper[4741]: I0929 20:57:01.739240 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:57:01 crc kubenswrapper[4741]: I0929 20:57:01.739281 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 20:57:01 crc kubenswrapper[4741]: I0929 20:57:01.740156 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9532bce3d47878ef326b645cc834aaa5500a0112cd2409697bc4431be4e30531"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 20:57:01 crc kubenswrapper[4741]: I0929 20:57:01.740205 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://9532bce3d47878ef326b645cc834aaa5500a0112cd2409697bc4431be4e30531" gracePeriod=600 Sep 29 20:57:02 crc kubenswrapper[4741]: I0929 20:57:02.461163 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="9532bce3d47878ef326b645cc834aaa5500a0112cd2409697bc4431be4e30531" exitCode=0 Sep 29 20:57:02 crc kubenswrapper[4741]: I0929 20:57:02.461219 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"9532bce3d47878ef326b645cc834aaa5500a0112cd2409697bc4431be4e30531"} Sep 29 20:57:02 crc kubenswrapper[4741]: I0929 20:57:02.461917 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97"} Sep 29 20:57:02 crc kubenswrapper[4741]: I0929 20:57:02.461948 4741 scope.go:117] "RemoveContainer" containerID="67b175830edb50b95a61957d8c9d5623aa554025cf94c8e53c5f8920542955c7" Sep 29 20:57:14 crc kubenswrapper[4741]: I0929 20:57:14.050114 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-rvkpf"] Sep 29 20:57:14 crc kubenswrapper[4741]: I0929 20:57:14.065569 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-rvkpf"] Sep 29 20:57:15 crc kubenswrapper[4741]: I0929 20:57:15.103370 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11060069-6fb7-4371-9be8-938a92c8269e" path="/var/lib/kubelet/pods/11060069-6fb7-4371-9be8-938a92c8269e/volumes" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.509047 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gxvj5"] Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.512895 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.531138 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gxvj5"] Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.652500 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-catalog-content\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.652662 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcm8d\" (UniqueName: \"kubernetes.io/projected/43c86a3a-a9d3-4633-98df-9c764ff2e545-kube-api-access-zcm8d\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.652796 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-utilities\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.755033 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-utilities\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.755235 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-catalog-content\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.755429 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcm8d\" (UniqueName: \"kubernetes.io/projected/43c86a3a-a9d3-4633-98df-9c764ff2e545-kube-api-access-zcm8d\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.756693 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-utilities\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.757127 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-catalog-content\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.776952 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcm8d\" (UniqueName: \"kubernetes.io/projected/43c86a3a-a9d3-4633-98df-9c764ff2e545-kube-api-access-zcm8d\") pod \"redhat-operators-gxvj5\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:21 crc kubenswrapper[4741]: I0929 20:57:21.850467 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:22 crc kubenswrapper[4741]: I0929 20:57:22.342833 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gxvj5"] Sep 29 20:57:22 crc kubenswrapper[4741]: I0929 20:57:22.712053 4741 generic.go:334] "Generic (PLEG): container finished" podID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerID="44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881" exitCode=0 Sep 29 20:57:22 crc kubenswrapper[4741]: I0929 20:57:22.712158 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxvj5" event={"ID":"43c86a3a-a9d3-4633-98df-9c764ff2e545","Type":"ContainerDied","Data":"44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881"} Sep 29 20:57:22 crc kubenswrapper[4741]: I0929 20:57:22.713212 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxvj5" event={"ID":"43c86a3a-a9d3-4633-98df-9c764ff2e545","Type":"ContainerStarted","Data":"d5677562f6493994729877049f87ff0e8e40ceee436cbb797ca2822147a648dc"} Sep 29 20:57:23 crc kubenswrapper[4741]: I0929 20:57:23.723664 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxvj5" event={"ID":"43c86a3a-a9d3-4633-98df-9c764ff2e545","Type":"ContainerStarted","Data":"75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a"} Sep 29 20:57:27 crc kubenswrapper[4741]: I0929 20:57:27.761318 4741 generic.go:334] "Generic (PLEG): container finished" podID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerID="75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a" exitCode=0 Sep 29 20:57:27 crc kubenswrapper[4741]: I0929 20:57:27.761359 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxvj5" event={"ID":"43c86a3a-a9d3-4633-98df-9c764ff2e545","Type":"ContainerDied","Data":"75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a"} Sep 29 20:57:28 crc kubenswrapper[4741]: I0929 20:57:28.777788 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxvj5" event={"ID":"43c86a3a-a9d3-4633-98df-9c764ff2e545","Type":"ContainerStarted","Data":"e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948"} Sep 29 20:57:28 crc kubenswrapper[4741]: I0929 20:57:28.798983 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gxvj5" podStartSLOduration=2.291236006 podStartE2EDuration="7.798970369s" podCreationTimestamp="2025-09-29 20:57:21 +0000 UTC" firstStartedPulling="2025-09-29 20:57:22.714089221 +0000 UTC m=+6484.361878553" lastFinishedPulling="2025-09-29 20:57:28.221823584 +0000 UTC m=+6489.869612916" observedRunningTime="2025-09-29 20:57:28.795230051 +0000 UTC m=+6490.443019383" watchObservedRunningTime="2025-09-29 20:57:28.798970369 +0000 UTC m=+6490.446759701" Sep 29 20:57:30 crc kubenswrapper[4741]: I0929 20:57:30.449188 4741 scope.go:117] "RemoveContainer" containerID="064b2dbb039ec7f838bf18a3a72321c366b352a397bacad477ea937782e2c581" Sep 29 20:57:30 crc kubenswrapper[4741]: I0929 20:57:30.477138 4741 scope.go:117] "RemoveContainer" containerID="0fbd7b96e2257f71ec10e9d0fbaf2b94ca5b1b36b6d1eff6f5afdaaf2bfcaa56" Sep 29 20:57:30 crc kubenswrapper[4741]: I0929 20:57:30.530050 4741 scope.go:117] "RemoveContainer" containerID="c1d008ea2aa0a500cbcc52c146bc49b0959bb30e5da2cd55abe927237a3389de" Sep 29 20:57:31 crc kubenswrapper[4741]: I0929 20:57:31.851448 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:31 crc kubenswrapper[4741]: I0929 20:57:31.851821 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:32 crc kubenswrapper[4741]: I0929 20:57:32.902896 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gxvj5" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="registry-server" probeResult="failure" output=< Sep 29 20:57:32 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 20:57:32 crc kubenswrapper[4741]: > Sep 29 20:57:41 crc kubenswrapper[4741]: I0929 20:57:41.925787 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:41 crc kubenswrapper[4741]: I0929 20:57:41.998086 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:42 crc kubenswrapper[4741]: I0929 20:57:42.166893 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gxvj5"] Sep 29 20:57:43 crc kubenswrapper[4741]: I0929 20:57:43.938391 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gxvj5" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="registry-server" containerID="cri-o://e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948" gracePeriod=2 Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.464619 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.575400 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-utilities\") pod \"43c86a3a-a9d3-4633-98df-9c764ff2e545\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.575509 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcm8d\" (UniqueName: \"kubernetes.io/projected/43c86a3a-a9d3-4633-98df-9c764ff2e545-kube-api-access-zcm8d\") pod \"43c86a3a-a9d3-4633-98df-9c764ff2e545\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.575604 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-catalog-content\") pod \"43c86a3a-a9d3-4633-98df-9c764ff2e545\" (UID: \"43c86a3a-a9d3-4633-98df-9c764ff2e545\") " Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.576195 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-utilities" (OuterVolumeSpecName: "utilities") pod "43c86a3a-a9d3-4633-98df-9c764ff2e545" (UID: "43c86a3a-a9d3-4633-98df-9c764ff2e545"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.577396 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.580959 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c86a3a-a9d3-4633-98df-9c764ff2e545-kube-api-access-zcm8d" (OuterVolumeSpecName: "kube-api-access-zcm8d") pod "43c86a3a-a9d3-4633-98df-9c764ff2e545" (UID: "43c86a3a-a9d3-4633-98df-9c764ff2e545"). InnerVolumeSpecName "kube-api-access-zcm8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.664550 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43c86a3a-a9d3-4633-98df-9c764ff2e545" (UID: "43c86a3a-a9d3-4633-98df-9c764ff2e545"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.679181 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcm8d\" (UniqueName: \"kubernetes.io/projected/43c86a3a-a9d3-4633-98df-9c764ff2e545-kube-api-access-zcm8d\") on node \"crc\" DevicePath \"\"" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.679396 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c86a3a-a9d3-4633-98df-9c764ff2e545-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.951244 4741 generic.go:334] "Generic (PLEG): container finished" podID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerID="e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948" exitCode=0 Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.951302 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxvj5" event={"ID":"43c86a3a-a9d3-4633-98df-9c764ff2e545","Type":"ContainerDied","Data":"e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948"} Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.951354 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxvj5" event={"ID":"43c86a3a-a9d3-4633-98df-9c764ff2e545","Type":"ContainerDied","Data":"d5677562f6493994729877049f87ff0e8e40ceee436cbb797ca2822147a648dc"} Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.951380 4741 scope.go:117] "RemoveContainer" containerID="e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.952375 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxvj5" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.983054 4741 scope.go:117] "RemoveContainer" containerID="75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a" Sep 29 20:57:44 crc kubenswrapper[4741]: I0929 20:57:44.988748 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gxvj5"] Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.000055 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gxvj5"] Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.008063 4741 scope.go:117] "RemoveContainer" containerID="44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881" Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.053371 4741 scope.go:117] "RemoveContainer" containerID="e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948" Sep 29 20:57:45 crc kubenswrapper[4741]: E0929 20:57:45.053839 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948\": container with ID starting with e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948 not found: ID does not exist" containerID="e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948" Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.053869 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948"} err="failed to get container status \"e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948\": rpc error: code = NotFound desc = could not find container \"e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948\": container with ID starting with e644127f68256a0b29322da54e91b0f6be632c4f8d8d3384cde09cf828472948 not found: ID does not exist" Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.053889 4741 scope.go:117] "RemoveContainer" containerID="75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a" Sep 29 20:57:45 crc kubenswrapper[4741]: E0929 20:57:45.054174 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a\": container with ID starting with 75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a not found: ID does not exist" containerID="75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a" Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.054192 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a"} err="failed to get container status \"75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a\": rpc error: code = NotFound desc = could not find container \"75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a\": container with ID starting with 75a1f70aea8828a5ac4fd76f7b7b9300168068d5d776575334e18d9da95a1d5a not found: ID does not exist" Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.054204 4741 scope.go:117] "RemoveContainer" containerID="44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881" Sep 29 20:57:45 crc kubenswrapper[4741]: E0929 20:57:45.054451 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881\": container with ID starting with 44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881 not found: ID does not exist" containerID="44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881" Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.054472 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881"} err="failed to get container status \"44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881\": rpc error: code = NotFound desc = could not find container \"44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881\": container with ID starting with 44ffbcf0e6c074504b3b864bb654ca4b0ea3cf689f3a45c6b6e254f461837881 not found: ID does not exist" Sep 29 20:57:45 crc kubenswrapper[4741]: I0929 20:57:45.100090 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" path="/var/lib/kubelet/pods/43c86a3a-a9d3-4633-98df-9c764ff2e545/volumes" Sep 29 20:59:31 crc kubenswrapper[4741]: I0929 20:59:31.738961 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 20:59:31 crc kubenswrapper[4741]: I0929 20:59:31.739549 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 20:59:42 crc kubenswrapper[4741]: I0929 20:59:42.035059 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-vz6r4"] Sep 29 20:59:42 crc kubenswrapper[4741]: I0929 20:59:42.044086 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-vz6r4"] Sep 29 20:59:43 crc kubenswrapper[4741]: I0929 20:59:43.122427 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="943c6d41-b403-45de-891e-16c5421ab9da" path="/var/lib/kubelet/pods/943c6d41-b403-45de-891e-16c5421ab9da/volumes" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.479898 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-psvxj"] Sep 29 20:59:46 crc kubenswrapper[4741]: E0929 20:59:46.480876 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="extract-content" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.480890 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="extract-content" Sep 29 20:59:46 crc kubenswrapper[4741]: E0929 20:59:46.480908 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="extract-utilities" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.480914 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="extract-utilities" Sep 29 20:59:46 crc kubenswrapper[4741]: E0929 20:59:46.480927 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="registry-server" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.480932 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="registry-server" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.481177 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c86a3a-a9d3-4633-98df-9c764ff2e545" containerName="registry-server" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.482808 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.497479 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-psvxj"] Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.565650 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-utilities\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.565710 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nb6q\" (UniqueName: \"kubernetes.io/projected/b874fb56-4776-406d-8545-326d6ea961ee-kube-api-access-4nb6q\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.565774 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-catalog-content\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.667458 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-utilities\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.667521 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nb6q\" (UniqueName: \"kubernetes.io/projected/b874fb56-4776-406d-8545-326d6ea961ee-kube-api-access-4nb6q\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.667601 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-catalog-content\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.668161 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-catalog-content\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.668191 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-utilities\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.687655 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nb6q\" (UniqueName: \"kubernetes.io/projected/b874fb56-4776-406d-8545-326d6ea961ee-kube-api-access-4nb6q\") pod \"certified-operators-psvxj\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:46 crc kubenswrapper[4741]: I0929 20:59:46.810872 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:47 crc kubenswrapper[4741]: I0929 20:59:47.296436 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-psvxj"] Sep 29 20:59:47 crc kubenswrapper[4741]: W0929 20:59:47.312336 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb874fb56_4776_406d_8545_326d6ea961ee.slice/crio-906ec17325ca5848921764e47b36cb9e364faeb1f2a5b23b9d325f50e61cce28 WatchSource:0}: Error finding container 906ec17325ca5848921764e47b36cb9e364faeb1f2a5b23b9d325f50e61cce28: Status 404 returned error can't find the container with id 906ec17325ca5848921764e47b36cb9e364faeb1f2a5b23b9d325f50e61cce28 Sep 29 20:59:48 crc kubenswrapper[4741]: I0929 20:59:48.279969 4741 generic.go:334] "Generic (PLEG): container finished" podID="b874fb56-4776-406d-8545-326d6ea961ee" containerID="7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902" exitCode=0 Sep 29 20:59:48 crc kubenswrapper[4741]: I0929 20:59:48.280029 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psvxj" event={"ID":"b874fb56-4776-406d-8545-326d6ea961ee","Type":"ContainerDied","Data":"7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902"} Sep 29 20:59:48 crc kubenswrapper[4741]: I0929 20:59:48.281175 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psvxj" event={"ID":"b874fb56-4776-406d-8545-326d6ea961ee","Type":"ContainerStarted","Data":"906ec17325ca5848921764e47b36cb9e364faeb1f2a5b23b9d325f50e61cce28"} Sep 29 20:59:48 crc kubenswrapper[4741]: I0929 20:59:48.282931 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 20:59:49 crc kubenswrapper[4741]: I0929 20:59:49.290165 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psvxj" event={"ID":"b874fb56-4776-406d-8545-326d6ea961ee","Type":"ContainerStarted","Data":"2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507"} Sep 29 20:59:50 crc kubenswrapper[4741]: I0929 20:59:50.300566 4741 generic.go:334] "Generic (PLEG): container finished" podID="b874fb56-4776-406d-8545-326d6ea961ee" containerID="2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507" exitCode=0 Sep 29 20:59:50 crc kubenswrapper[4741]: I0929 20:59:50.301023 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psvxj" event={"ID":"b874fb56-4776-406d-8545-326d6ea961ee","Type":"ContainerDied","Data":"2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507"} Sep 29 20:59:51 crc kubenswrapper[4741]: I0929 20:59:51.311451 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psvxj" event={"ID":"b874fb56-4776-406d-8545-326d6ea961ee","Type":"ContainerStarted","Data":"9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d"} Sep 29 20:59:51 crc kubenswrapper[4741]: I0929 20:59:51.328327 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-psvxj" podStartSLOduration=2.88619324 podStartE2EDuration="5.328310614s" podCreationTimestamp="2025-09-29 20:59:46 +0000 UTC" firstStartedPulling="2025-09-29 20:59:48.282693145 +0000 UTC m=+6629.930482477" lastFinishedPulling="2025-09-29 20:59:50.724810519 +0000 UTC m=+6632.372599851" observedRunningTime="2025-09-29 20:59:51.326580729 +0000 UTC m=+6632.974370071" watchObservedRunningTime="2025-09-29 20:59:51.328310614 +0000 UTC m=+6632.976099946" Sep 29 20:59:52 crc kubenswrapper[4741]: I0929 20:59:52.025485 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-a79d-account-create-jm4qm"] Sep 29 20:59:52 crc kubenswrapper[4741]: I0929 20:59:52.032910 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-a79d-account-create-jm4qm"] Sep 29 20:59:53 crc kubenswrapper[4741]: I0929 20:59:53.106000 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aee1d090-e1c4-47ce-8bc9-03fa600db9fa" path="/var/lib/kubelet/pods/aee1d090-e1c4-47ce-8bc9-03fa600db9fa/volumes" Sep 29 20:59:56 crc kubenswrapper[4741]: I0929 20:59:56.811307 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:56 crc kubenswrapper[4741]: I0929 20:59:56.812460 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:56 crc kubenswrapper[4741]: I0929 20:59:56.884953 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:57 crc kubenswrapper[4741]: I0929 20:59:57.432461 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:57 crc kubenswrapper[4741]: I0929 20:59:57.492273 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-psvxj"] Sep 29 20:59:59 crc kubenswrapper[4741]: I0929 20:59:59.383076 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-psvxj" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="registry-server" containerID="cri-o://9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d" gracePeriod=2 Sep 29 20:59:59 crc kubenswrapper[4741]: I0929 20:59:59.858923 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 20:59:59 crc kubenswrapper[4741]: I0929 20:59:59.954257 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-utilities\") pod \"b874fb56-4776-406d-8545-326d6ea961ee\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " Sep 29 20:59:59 crc kubenswrapper[4741]: I0929 20:59:59.954311 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-catalog-content\") pod \"b874fb56-4776-406d-8545-326d6ea961ee\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " Sep 29 20:59:59 crc kubenswrapper[4741]: I0929 20:59:59.954434 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nb6q\" (UniqueName: \"kubernetes.io/projected/b874fb56-4776-406d-8545-326d6ea961ee-kube-api-access-4nb6q\") pod \"b874fb56-4776-406d-8545-326d6ea961ee\" (UID: \"b874fb56-4776-406d-8545-326d6ea961ee\") " Sep 29 20:59:59 crc kubenswrapper[4741]: I0929 20:59:59.956313 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-utilities" (OuterVolumeSpecName: "utilities") pod "b874fb56-4776-406d-8545-326d6ea961ee" (UID: "b874fb56-4776-406d-8545-326d6ea961ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 20:59:59 crc kubenswrapper[4741]: I0929 20:59:59.960612 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b874fb56-4776-406d-8545-326d6ea961ee-kube-api-access-4nb6q" (OuterVolumeSpecName: "kube-api-access-4nb6q") pod "b874fb56-4776-406d-8545-326d6ea961ee" (UID: "b874fb56-4776-406d-8545-326d6ea961ee"). InnerVolumeSpecName "kube-api-access-4nb6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.055942 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nb6q\" (UniqueName: \"kubernetes.io/projected/b874fb56-4776-406d-8545-326d6ea961ee-kube-api-access-4nb6q\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.055971 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.166358 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz"] Sep 29 21:00:00 crc kubenswrapper[4741]: E0929 21:00:00.167029 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="extract-utilities" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.167082 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="extract-utilities" Sep 29 21:00:00 crc kubenswrapper[4741]: E0929 21:00:00.167108 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="extract-content" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.167115 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="extract-content" Sep 29 21:00:00 crc kubenswrapper[4741]: E0929 21:00:00.167156 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="registry-server" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.167165 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="registry-server" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.167462 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b874fb56-4776-406d-8545-326d6ea961ee" containerName="registry-server" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.168455 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.171069 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.171076 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.176373 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz"] Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.263688 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlqtf\" (UniqueName: \"kubernetes.io/projected/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-kube-api-access-wlqtf\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.263845 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-secret-volume\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.263962 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-config-volume\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.366300 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-config-volume\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.366371 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlqtf\" (UniqueName: \"kubernetes.io/projected/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-kube-api-access-wlqtf\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.366476 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-secret-volume\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.367091 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-config-volume\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.370227 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-secret-volume\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.383286 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlqtf\" (UniqueName: \"kubernetes.io/projected/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-kube-api-access-wlqtf\") pod \"collect-profiles-29319660-nl4kz\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.393417 4741 generic.go:334] "Generic (PLEG): container finished" podID="b874fb56-4776-406d-8545-326d6ea961ee" containerID="9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d" exitCode=0 Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.393462 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psvxj" event={"ID":"b874fb56-4776-406d-8545-326d6ea961ee","Type":"ContainerDied","Data":"9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d"} Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.393508 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psvxj" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.393528 4741 scope.go:117] "RemoveContainer" containerID="9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.393513 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psvxj" event={"ID":"b874fb56-4776-406d-8545-326d6ea961ee","Type":"ContainerDied","Data":"906ec17325ca5848921764e47b36cb9e364faeb1f2a5b23b9d325f50e61cce28"} Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.435071 4741 scope.go:117] "RemoveContainer" containerID="2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.462528 4741 scope.go:117] "RemoveContainer" containerID="7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.481805 4741 scope.go:117] "RemoveContainer" containerID="9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d" Sep 29 21:00:00 crc kubenswrapper[4741]: E0929 21:00:00.482223 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d\": container with ID starting with 9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d not found: ID does not exist" containerID="9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.482261 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d"} err="failed to get container status \"9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d\": rpc error: code = NotFound desc = could not find container \"9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d\": container with ID starting with 9fc3a2cb50baa9d09938e08e9986d81227012ae77d259a3e3d7e2a4d7c8fe75d not found: ID does not exist" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.482286 4741 scope.go:117] "RemoveContainer" containerID="2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507" Sep 29 21:00:00 crc kubenswrapper[4741]: E0929 21:00:00.482810 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507\": container with ID starting with 2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507 not found: ID does not exist" containerID="2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.482849 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507"} err="failed to get container status \"2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507\": rpc error: code = NotFound desc = could not find container \"2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507\": container with ID starting with 2a16fbd5053efc3e40094946220bc91a68af2e43e9cff6c82aa7ff9974dca507 not found: ID does not exist" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.482874 4741 scope.go:117] "RemoveContainer" containerID="7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902" Sep 29 21:00:00 crc kubenswrapper[4741]: E0929 21:00:00.483146 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902\": container with ID starting with 7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902 not found: ID does not exist" containerID="7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.483171 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902"} err="failed to get container status \"7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902\": rpc error: code = NotFound desc = could not find container \"7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902\": container with ID starting with 7183c7892cef200b91358d564ec62650a7bd017c92d322ed977fc0b0b9da3902 not found: ID does not exist" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.527035 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b874fb56-4776-406d-8545-326d6ea961ee" (UID: "b874fb56-4776-406d-8545-326d6ea961ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.537157 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.570972 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b874fb56-4776-406d-8545-326d6ea961ee-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.767197 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-psvxj"] Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.787145 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-psvxj"] Sep 29 21:00:00 crc kubenswrapper[4741]: I0929 21:00:00.999296 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz"] Sep 29 21:00:01 crc kubenswrapper[4741]: I0929 21:00:01.102719 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b874fb56-4776-406d-8545-326d6ea961ee" path="/var/lib/kubelet/pods/b874fb56-4776-406d-8545-326d6ea961ee/volumes" Sep 29 21:00:01 crc kubenswrapper[4741]: I0929 21:00:01.406529 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" event={"ID":"318cac1a-c05b-4e3c-a379-cbcbbf8cd071","Type":"ContainerStarted","Data":"c7a38977ae13f188cb856e63905e8deb59e17ee82118871d26e8676cd6658eca"} Sep 29 21:00:01 crc kubenswrapper[4741]: I0929 21:00:01.406933 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" event={"ID":"318cac1a-c05b-4e3c-a379-cbcbbf8cd071","Type":"ContainerStarted","Data":"3cbea00d6bd86de2dd2a8957b5f2223ed8edc8187b9d59c71a2d68785c145000"} Sep 29 21:00:01 crc kubenswrapper[4741]: I0929 21:00:01.424094 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" podStartSLOduration=1.424075119 podStartE2EDuration="1.424075119s" podCreationTimestamp="2025-09-29 21:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:00:01.420197627 +0000 UTC m=+6643.067986959" watchObservedRunningTime="2025-09-29 21:00:01.424075119 +0000 UTC m=+6643.071864451" Sep 29 21:00:01 crc kubenswrapper[4741]: I0929 21:00:01.739223 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:00:01 crc kubenswrapper[4741]: I0929 21:00:01.739276 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:00:02 crc kubenswrapper[4741]: I0929 21:00:02.416590 4741 generic.go:334] "Generic (PLEG): container finished" podID="318cac1a-c05b-4e3c-a379-cbcbbf8cd071" containerID="c7a38977ae13f188cb856e63905e8deb59e17ee82118871d26e8676cd6658eca" exitCode=0 Sep 29 21:00:02 crc kubenswrapper[4741]: I0929 21:00:02.416759 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" event={"ID":"318cac1a-c05b-4e3c-a379-cbcbbf8cd071","Type":"ContainerDied","Data":"c7a38977ae13f188cb856e63905e8deb59e17ee82118871d26e8676cd6658eca"} Sep 29 21:00:03 crc kubenswrapper[4741]: I0929 21:00:03.854729 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.047212 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-secret-volume\") pod \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.047555 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-config-volume\") pod \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.047609 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlqtf\" (UniqueName: \"kubernetes.io/projected/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-kube-api-access-wlqtf\") pod \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\" (UID: \"318cac1a-c05b-4e3c-a379-cbcbbf8cd071\") " Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.047998 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-config-volume" (OuterVolumeSpecName: "config-volume") pod "318cac1a-c05b-4e3c-a379-cbcbbf8cd071" (UID: "318cac1a-c05b-4e3c-a379-cbcbbf8cd071"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.048300 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.062826 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "318cac1a-c05b-4e3c-a379-cbcbbf8cd071" (UID: "318cac1a-c05b-4e3c-a379-cbcbbf8cd071"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.064630 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-kube-api-access-wlqtf" (OuterVolumeSpecName: "kube-api-access-wlqtf") pod "318cac1a-c05b-4e3c-a379-cbcbbf8cd071" (UID: "318cac1a-c05b-4e3c-a379-cbcbbf8cd071"). InnerVolumeSpecName "kube-api-access-wlqtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.150868 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlqtf\" (UniqueName: \"kubernetes.io/projected/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-kube-api-access-wlqtf\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.150898 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/318cac1a-c05b-4e3c-a379-cbcbbf8cd071-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.448116 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" event={"ID":"318cac1a-c05b-4e3c-a379-cbcbbf8cd071","Type":"ContainerDied","Data":"3cbea00d6bd86de2dd2a8957b5f2223ed8edc8187b9d59c71a2d68785c145000"} Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.448161 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cbea00d6bd86de2dd2a8957b5f2223ed8edc8187b9d59c71a2d68785c145000" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.448183 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz" Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.491473 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s"] Sep 29 21:00:04 crc kubenswrapper[4741]: I0929 21:00:04.500156 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319615-ftp9s"] Sep 29 21:00:05 crc kubenswrapper[4741]: I0929 21:00:05.027523 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-d5dz9"] Sep 29 21:00:05 crc kubenswrapper[4741]: I0929 21:00:05.035423 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-d5dz9"] Sep 29 21:00:05 crc kubenswrapper[4741]: I0929 21:00:05.103407 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302998d5-a7fc-458f-93e4-6e1893537638" path="/var/lib/kubelet/pods/302998d5-a7fc-458f-93e4-6e1893537638/volumes" Sep 29 21:00:05 crc kubenswrapper[4741]: I0929 21:00:05.104350 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7fbd8d6-0243-4044-b907-8bf0448d58f2" path="/var/lib/kubelet/pods/c7fbd8d6-0243-4044-b907-8bf0448d58f2/volumes" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.208542 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wdzxv"] Sep 29 21:00:12 crc kubenswrapper[4741]: E0929 21:00:12.209459 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318cac1a-c05b-4e3c-a379-cbcbbf8cd071" containerName="collect-profiles" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.209471 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="318cac1a-c05b-4e3c-a379-cbcbbf8cd071" containerName="collect-profiles" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.209679 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="318cac1a-c05b-4e3c-a379-cbcbbf8cd071" containerName="collect-profiles" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.211326 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.230436 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdzxv"] Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.239669 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-catalog-content\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.239752 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-utilities\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.239899 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xgqn\" (UniqueName: \"kubernetes.io/projected/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-kube-api-access-7xgqn\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.341672 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-catalog-content\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.341830 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-utilities\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.341912 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xgqn\" (UniqueName: \"kubernetes.io/projected/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-kube-api-access-7xgqn\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.342241 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-catalog-content\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.342294 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-utilities\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.365053 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xgqn\" (UniqueName: \"kubernetes.io/projected/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-kube-api-access-7xgqn\") pod \"redhat-marketplace-wdzxv\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:12 crc kubenswrapper[4741]: I0929 21:00:12.533543 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:13 crc kubenswrapper[4741]: I0929 21:00:13.034440 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdzxv"] Sep 29 21:00:13 crc kubenswrapper[4741]: I0929 21:00:13.543471 4741 generic.go:334] "Generic (PLEG): container finished" podID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerID="df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a" exitCode=0 Sep 29 21:00:13 crc kubenswrapper[4741]: I0929 21:00:13.543533 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdzxv" event={"ID":"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74","Type":"ContainerDied","Data":"df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a"} Sep 29 21:00:13 crc kubenswrapper[4741]: I0929 21:00:13.543606 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdzxv" event={"ID":"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74","Type":"ContainerStarted","Data":"8d8649e9eed9f37ce5ac1edd87f4c850c141a28bb35acf25cb726025bef747af"} Sep 29 21:00:14 crc kubenswrapper[4741]: I0929 21:00:14.553086 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdzxv" event={"ID":"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74","Type":"ContainerStarted","Data":"154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323"} Sep 29 21:00:15 crc kubenswrapper[4741]: E0929 21:00:15.009475 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9e1ca7f_7da3_424e_aa63_efcf0c4abd74.slice/crio-154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323.scope\": RecentStats: unable to find data in memory cache]" Sep 29 21:00:15 crc kubenswrapper[4741]: I0929 21:00:15.568712 4741 generic.go:334] "Generic (PLEG): container finished" podID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerID="154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323" exitCode=0 Sep 29 21:00:15 crc kubenswrapper[4741]: I0929 21:00:15.568763 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdzxv" event={"ID":"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74","Type":"ContainerDied","Data":"154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323"} Sep 29 21:00:16 crc kubenswrapper[4741]: I0929 21:00:16.580365 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdzxv" event={"ID":"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74","Type":"ContainerStarted","Data":"186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152"} Sep 29 21:00:16 crc kubenswrapper[4741]: I0929 21:00:16.606775 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wdzxv" podStartSLOduration=2.157781354 podStartE2EDuration="4.606746084s" podCreationTimestamp="2025-09-29 21:00:12 +0000 UTC" firstStartedPulling="2025-09-29 21:00:13.545868185 +0000 UTC m=+6655.193657517" lastFinishedPulling="2025-09-29 21:00:15.994832915 +0000 UTC m=+6657.642622247" observedRunningTime="2025-09-29 21:00:16.596316556 +0000 UTC m=+6658.244105898" watchObservedRunningTime="2025-09-29 21:00:16.606746084 +0000 UTC m=+6658.254535436" Sep 29 21:00:22 crc kubenswrapper[4741]: I0929 21:00:22.534035 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:22 crc kubenswrapper[4741]: I0929 21:00:22.534732 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:22 crc kubenswrapper[4741]: I0929 21:00:22.584282 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:22 crc kubenswrapper[4741]: I0929 21:00:22.687578 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:22 crc kubenswrapper[4741]: I0929 21:00:22.822671 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdzxv"] Sep 29 21:00:24 crc kubenswrapper[4741]: I0929 21:00:24.651472 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wdzxv" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="registry-server" containerID="cri-o://186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152" gracePeriod=2 Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.221891 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.333817 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-catalog-content\") pod \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.334325 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xgqn\" (UniqueName: \"kubernetes.io/projected/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-kube-api-access-7xgqn\") pod \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.334382 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-utilities\") pod \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\" (UID: \"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74\") " Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.335142 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-utilities" (OuterVolumeSpecName: "utilities") pod "a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" (UID: "a9e1ca7f-7da3-424e-aa63-efcf0c4abd74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.335429 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.339785 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-kube-api-access-7xgqn" (OuterVolumeSpecName: "kube-api-access-7xgqn") pod "a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" (UID: "a9e1ca7f-7da3-424e-aa63-efcf0c4abd74"). InnerVolumeSpecName "kube-api-access-7xgqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.346166 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" (UID: "a9e1ca7f-7da3-424e-aa63-efcf0c4abd74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.438100 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xgqn\" (UniqueName: \"kubernetes.io/projected/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-kube-api-access-7xgqn\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.438144 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.662536 4741 generic.go:334] "Generic (PLEG): container finished" podID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerID="186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152" exitCode=0 Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.662574 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdzxv" event={"ID":"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74","Type":"ContainerDied","Data":"186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152"} Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.662607 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdzxv" event={"ID":"a9e1ca7f-7da3-424e-aa63-efcf0c4abd74","Type":"ContainerDied","Data":"8d8649e9eed9f37ce5ac1edd87f4c850c141a28bb35acf25cb726025bef747af"} Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.662625 4741 scope.go:117] "RemoveContainer" containerID="186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.663556 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdzxv" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.682671 4741 scope.go:117] "RemoveContainer" containerID="154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.711600 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdzxv"] Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.719071 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdzxv"] Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.722814 4741 scope.go:117] "RemoveContainer" containerID="df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.754350 4741 scope.go:117] "RemoveContainer" containerID="186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152" Sep 29 21:00:25 crc kubenswrapper[4741]: E0929 21:00:25.754773 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152\": container with ID starting with 186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152 not found: ID does not exist" containerID="186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.754841 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152"} err="failed to get container status \"186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152\": rpc error: code = NotFound desc = could not find container \"186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152\": container with ID starting with 186936b2fb87c5873f31de40d7f6092b8b95fed3bc6e9d89d48566485bb60152 not found: ID does not exist" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.754869 4741 scope.go:117] "RemoveContainer" containerID="154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323" Sep 29 21:00:25 crc kubenswrapper[4741]: E0929 21:00:25.755215 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323\": container with ID starting with 154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323 not found: ID does not exist" containerID="154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.755277 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323"} err="failed to get container status \"154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323\": rpc error: code = NotFound desc = could not find container \"154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323\": container with ID starting with 154d1e85397054c9bfd4d35c1ee0d3ed802dca9d7816bcbf44f35384bf56c323 not found: ID does not exist" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.755314 4741 scope.go:117] "RemoveContainer" containerID="df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a" Sep 29 21:00:25 crc kubenswrapper[4741]: E0929 21:00:25.755667 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a\": container with ID starting with df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a not found: ID does not exist" containerID="df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a" Sep 29 21:00:25 crc kubenswrapper[4741]: I0929 21:00:25.755690 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a"} err="failed to get container status \"df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a\": rpc error: code = NotFound desc = could not find container \"df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a\": container with ID starting with df9ffd56c6d207247dc9884ce00efd5c9b0fa907653a126b9b2c74744051577a not found: ID does not exist" Sep 29 21:00:27 crc kubenswrapper[4741]: I0929 21:00:27.100073 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" path="/var/lib/kubelet/pods/a9e1ca7f-7da3-424e-aa63-efcf0c4abd74/volumes" Sep 29 21:00:30 crc kubenswrapper[4741]: I0929 21:00:30.695049 4741 scope.go:117] "RemoveContainer" containerID="35338d1b61e585850d495e402eb13ffb9f6a5dec35f6a6c2a4c281db31552a5a" Sep 29 21:00:30 crc kubenswrapper[4741]: I0929 21:00:30.735183 4741 scope.go:117] "RemoveContainer" containerID="7c37a22716f28cff2a47266b8c405cc1ae739bfe4d51105630a8156a6be79cab" Sep 29 21:00:30 crc kubenswrapper[4741]: I0929 21:00:30.797586 4741 scope.go:117] "RemoveContainer" containerID="e423be380e915d310610a86192b1dd47582511ea8ce7d2d85805bd49737ac3c9" Sep 29 21:00:30 crc kubenswrapper[4741]: I0929 21:00:30.881595 4741 scope.go:117] "RemoveContainer" containerID="b86ce3936d8731243ff7b70216082259a6dbdf899cc96c3e442cd4beb0bbb62b" Sep 29 21:00:30 crc kubenswrapper[4741]: I0929 21:00:30.941530 4741 scope.go:117] "RemoveContainer" containerID="26686b3dbab6234575f00d623fbe300fe9c2d7e583e9ba940295cf316fc259c3" Sep 29 21:00:30 crc kubenswrapper[4741]: I0929 21:00:30.971130 4741 scope.go:117] "RemoveContainer" containerID="647d5e5a59c9b4b387af8969c28ddf5211f8a0a24b30831fcd87fadbfe9d7004" Sep 29 21:00:31 crc kubenswrapper[4741]: I0929 21:00:31.738742 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:00:31 crc kubenswrapper[4741]: I0929 21:00:31.738803 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:00:31 crc kubenswrapper[4741]: I0929 21:00:31.738841 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:00:31 crc kubenswrapper[4741]: I0929 21:00:31.739669 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:00:31 crc kubenswrapper[4741]: I0929 21:00:31.739734 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" gracePeriod=600 Sep 29 21:00:31 crc kubenswrapper[4741]: E0929 21:00:31.892509 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:00:32 crc kubenswrapper[4741]: I0929 21:00:32.751801 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" exitCode=0 Sep 29 21:00:32 crc kubenswrapper[4741]: I0929 21:00:32.751861 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97"} Sep 29 21:00:32 crc kubenswrapper[4741]: I0929 21:00:32.752189 4741 scope.go:117] "RemoveContainer" containerID="9532bce3d47878ef326b645cc834aaa5500a0112cd2409697bc4431be4e30531" Sep 29 21:00:32 crc kubenswrapper[4741]: I0929 21:00:32.753340 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:00:32 crc kubenswrapper[4741]: E0929 21:00:32.753962 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:00:45 crc kubenswrapper[4741]: I0929 21:00:45.086165 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:00:45 crc kubenswrapper[4741]: E0929 21:00:45.086964 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.086575 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:01:00 crc kubenswrapper[4741]: E0929 21:01:00.087566 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.153168 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29319661-lblwq"] Sep 29 21:01:00 crc kubenswrapper[4741]: E0929 21:01:00.154178 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="registry-server" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.154196 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="registry-server" Sep 29 21:01:00 crc kubenswrapper[4741]: E0929 21:01:00.154303 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="extract-utilities" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.154316 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="extract-utilities" Sep 29 21:01:00 crc kubenswrapper[4741]: E0929 21:01:00.154330 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="extract-content" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.154359 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="extract-content" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.154740 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9e1ca7f-7da3-424e-aa63-efcf0c4abd74" containerName="registry-server" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.155740 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.163370 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29319661-lblwq"] Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.240226 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-fernet-keys\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.240346 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f69wd\" (UniqueName: \"kubernetes.io/projected/51eef39d-57c6-4886-9b55-3378d8527072-kube-api-access-f69wd\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.240598 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-combined-ca-bundle\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.240714 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-config-data\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.344071 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-combined-ca-bundle\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.344635 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-config-data\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.344920 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-fernet-keys\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.345004 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f69wd\" (UniqueName: \"kubernetes.io/projected/51eef39d-57c6-4886-9b55-3378d8527072-kube-api-access-f69wd\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.349978 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-combined-ca-bundle\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.350067 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-config-data\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.351239 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-fernet-keys\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.360441 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f69wd\" (UniqueName: \"kubernetes.io/projected/51eef39d-57c6-4886-9b55-3378d8527072-kube-api-access-f69wd\") pod \"keystone-cron-29319661-lblwq\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.502978 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:00 crc kubenswrapper[4741]: I0929 21:01:00.985267 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29319661-lblwq"] Sep 29 21:01:01 crc kubenswrapper[4741]: I0929 21:01:01.058276 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319661-lblwq" event={"ID":"51eef39d-57c6-4886-9b55-3378d8527072","Type":"ContainerStarted","Data":"86391b6e3d420bfc0c4955e2239d0bf18321ead7fb14a7748481f263b0a3e61f"} Sep 29 21:01:02 crc kubenswrapper[4741]: I0929 21:01:02.071955 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319661-lblwq" event={"ID":"51eef39d-57c6-4886-9b55-3378d8527072","Type":"ContainerStarted","Data":"9c8d769530e7f0f821ddb0e3720f5babab5f259baedb01e900909764dfc59dd4"} Sep 29 21:01:02 crc kubenswrapper[4741]: I0929 21:01:02.095099 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29319661-lblwq" podStartSLOduration=2.095077493 podStartE2EDuration="2.095077493s" podCreationTimestamp="2025-09-29 21:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:01:02.088428423 +0000 UTC m=+6703.736217785" watchObservedRunningTime="2025-09-29 21:01:02.095077493 +0000 UTC m=+6703.742866835" Sep 29 21:01:04 crc kubenswrapper[4741]: I0929 21:01:04.097293 4741 generic.go:334] "Generic (PLEG): container finished" podID="51eef39d-57c6-4886-9b55-3378d8527072" containerID="9c8d769530e7f0f821ddb0e3720f5babab5f259baedb01e900909764dfc59dd4" exitCode=0 Sep 29 21:01:04 crc kubenswrapper[4741]: I0929 21:01:04.097341 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319661-lblwq" event={"ID":"51eef39d-57c6-4886-9b55-3378d8527072","Type":"ContainerDied","Data":"9c8d769530e7f0f821ddb0e3720f5babab5f259baedb01e900909764dfc59dd4"} Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.492351 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.507635 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-fernet-keys\") pod \"51eef39d-57c6-4886-9b55-3378d8527072\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.507776 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f69wd\" (UniqueName: \"kubernetes.io/projected/51eef39d-57c6-4886-9b55-3378d8527072-kube-api-access-f69wd\") pod \"51eef39d-57c6-4886-9b55-3378d8527072\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.507978 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-config-data\") pod \"51eef39d-57c6-4886-9b55-3378d8527072\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.508172 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-combined-ca-bundle\") pod \"51eef39d-57c6-4886-9b55-3378d8527072\" (UID: \"51eef39d-57c6-4886-9b55-3378d8527072\") " Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.514534 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51eef39d-57c6-4886-9b55-3378d8527072-kube-api-access-f69wd" (OuterVolumeSpecName: "kube-api-access-f69wd") pod "51eef39d-57c6-4886-9b55-3378d8527072" (UID: "51eef39d-57c6-4886-9b55-3378d8527072"). InnerVolumeSpecName "kube-api-access-f69wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.516559 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "51eef39d-57c6-4886-9b55-3378d8527072" (UID: "51eef39d-57c6-4886-9b55-3378d8527072"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.545446 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51eef39d-57c6-4886-9b55-3378d8527072" (UID: "51eef39d-57c6-4886-9b55-3378d8527072"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.566412 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-config-data" (OuterVolumeSpecName: "config-data") pod "51eef39d-57c6-4886-9b55-3378d8527072" (UID: "51eef39d-57c6-4886-9b55-3378d8527072"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.611114 4741 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.611148 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f69wd\" (UniqueName: \"kubernetes.io/projected/51eef39d-57c6-4886-9b55-3378d8527072-kube-api-access-f69wd\") on node \"crc\" DevicePath \"\"" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.611158 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 21:01:05 crc kubenswrapper[4741]: I0929 21:01:05.611167 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51eef39d-57c6-4886-9b55-3378d8527072-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:01:06 crc kubenswrapper[4741]: I0929 21:01:06.114233 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319661-lblwq" event={"ID":"51eef39d-57c6-4886-9b55-3378d8527072","Type":"ContainerDied","Data":"86391b6e3d420bfc0c4955e2239d0bf18321ead7fb14a7748481f263b0a3e61f"} Sep 29 21:01:06 crc kubenswrapper[4741]: I0929 21:01:06.114272 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86391b6e3d420bfc0c4955e2239d0bf18321ead7fb14a7748481f263b0a3e61f" Sep 29 21:01:06 crc kubenswrapper[4741]: I0929 21:01:06.114327 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319661-lblwq" Sep 29 21:01:12 crc kubenswrapper[4741]: I0929 21:01:12.086329 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:01:12 crc kubenswrapper[4741]: E0929 21:01:12.087358 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:01:26 crc kubenswrapper[4741]: I0929 21:01:26.086149 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:01:26 crc kubenswrapper[4741]: E0929 21:01:26.087031 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:01:41 crc kubenswrapper[4741]: I0929 21:01:41.087761 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:01:41 crc kubenswrapper[4741]: E0929 21:01:41.088982 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:01:54 crc kubenswrapper[4741]: I0929 21:01:54.086489 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:01:54 crc kubenswrapper[4741]: E0929 21:01:54.087153 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:02:05 crc kubenswrapper[4741]: I0929 21:02:05.086994 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:02:05 crc kubenswrapper[4741]: E0929 21:02:05.087767 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:02:06 crc kubenswrapper[4741]: I0929 21:02:06.056921 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-tldwl"] Sep 29 21:02:06 crc kubenswrapper[4741]: I0929 21:02:06.067581 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-tldwl"] Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.099219 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73a732d9-cfa4-4cb0-b2a8-2643777b89fb" path="/var/lib/kubelet/pods/73a732d9-cfa4-4cb0-b2a8-2643777b89fb/volumes" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.117499 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cgvcb"] Sep 29 21:02:07 crc kubenswrapper[4741]: E0929 21:02:07.118009 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51eef39d-57c6-4886-9b55-3378d8527072" containerName="keystone-cron" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.118032 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="51eef39d-57c6-4886-9b55-3378d8527072" containerName="keystone-cron" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.118217 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="51eef39d-57c6-4886-9b55-3378d8527072" containerName="keystone-cron" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.119695 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.146617 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgvcb"] Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.255081 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-utilities\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.255258 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-catalog-content\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.255301 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8rzd\" (UniqueName: \"kubernetes.io/projected/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-kube-api-access-j8rzd\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.357053 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-utilities\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.357155 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-catalog-content\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.357181 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8rzd\" (UniqueName: \"kubernetes.io/projected/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-kube-api-access-j8rzd\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.357698 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-utilities\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.357945 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-catalog-content\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.378236 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8rzd\" (UniqueName: \"kubernetes.io/projected/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-kube-api-access-j8rzd\") pod \"community-operators-cgvcb\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:07 crc kubenswrapper[4741]: I0929 21:02:07.460456 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:08 crc kubenswrapper[4741]: I0929 21:02:08.004089 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgvcb"] Sep 29 21:02:08 crc kubenswrapper[4741]: I0929 21:02:08.774599 4741 generic.go:334] "Generic (PLEG): container finished" podID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerID="27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8" exitCode=0 Sep 29 21:02:08 crc kubenswrapper[4741]: I0929 21:02:08.774648 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgvcb" event={"ID":"080e4b8e-cfb9-4bf1-bb53-be9c18606e78","Type":"ContainerDied","Data":"27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8"} Sep 29 21:02:08 crc kubenswrapper[4741]: I0929 21:02:08.775240 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgvcb" event={"ID":"080e4b8e-cfb9-4bf1-bb53-be9c18606e78","Type":"ContainerStarted","Data":"a1717e68952d297a90ffb1d3aeca8910f5f162da0a58217d6372032bf0ac30c6"} Sep 29 21:02:10 crc kubenswrapper[4741]: I0929 21:02:10.797216 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgvcb" event={"ID":"080e4b8e-cfb9-4bf1-bb53-be9c18606e78","Type":"ContainerStarted","Data":"c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60"} Sep 29 21:02:11 crc kubenswrapper[4741]: I0929 21:02:11.815381 4741 generic.go:334] "Generic (PLEG): container finished" podID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerID="c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60" exitCode=0 Sep 29 21:02:11 crc kubenswrapper[4741]: I0929 21:02:11.815467 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgvcb" event={"ID":"080e4b8e-cfb9-4bf1-bb53-be9c18606e78","Type":"ContainerDied","Data":"c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60"} Sep 29 21:02:12 crc kubenswrapper[4741]: I0929 21:02:12.827520 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgvcb" event={"ID":"080e4b8e-cfb9-4bf1-bb53-be9c18606e78","Type":"ContainerStarted","Data":"ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3"} Sep 29 21:02:12 crc kubenswrapper[4741]: I0929 21:02:12.850556 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cgvcb" podStartSLOduration=2.270075565 podStartE2EDuration="5.850533587s" podCreationTimestamp="2025-09-29 21:02:07 +0000 UTC" firstStartedPulling="2025-09-29 21:02:08.778004429 +0000 UTC m=+6770.425793761" lastFinishedPulling="2025-09-29 21:02:12.358462411 +0000 UTC m=+6774.006251783" observedRunningTime="2025-09-29 21:02:12.841960107 +0000 UTC m=+6774.489749449" watchObservedRunningTime="2025-09-29 21:02:12.850533587 +0000 UTC m=+6774.498322939" Sep 29 21:02:16 crc kubenswrapper[4741]: I0929 21:02:16.034208 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-a3d6-account-create-x59t8"] Sep 29 21:02:16 crc kubenswrapper[4741]: I0929 21:02:16.042867 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-a3d6-account-create-x59t8"] Sep 29 21:02:16 crc kubenswrapper[4741]: I0929 21:02:16.086211 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:02:16 crc kubenswrapper[4741]: E0929 21:02:16.086515 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:02:17 crc kubenswrapper[4741]: I0929 21:02:17.097068 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ed0b05b-c120-420f-907a-44aa6c04f7f8" path="/var/lib/kubelet/pods/8ed0b05b-c120-420f-907a-44aa6c04f7f8/volumes" Sep 29 21:02:17 crc kubenswrapper[4741]: I0929 21:02:17.460792 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:17 crc kubenswrapper[4741]: I0929 21:02:17.460839 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:17 crc kubenswrapper[4741]: I0929 21:02:17.507607 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:17 crc kubenswrapper[4741]: I0929 21:02:17.943257 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:17 crc kubenswrapper[4741]: I0929 21:02:17.989190 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgvcb"] Sep 29 21:02:19 crc kubenswrapper[4741]: I0929 21:02:19.899629 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cgvcb" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="registry-server" containerID="cri-o://ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3" gracePeriod=2 Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.395822 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.574446 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-utilities\") pod \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.574632 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-catalog-content\") pod \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.574674 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8rzd\" (UniqueName: \"kubernetes.io/projected/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-kube-api-access-j8rzd\") pod \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\" (UID: \"080e4b8e-cfb9-4bf1-bb53-be9c18606e78\") " Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.575685 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-utilities" (OuterVolumeSpecName: "utilities") pod "080e4b8e-cfb9-4bf1-bb53-be9c18606e78" (UID: "080e4b8e-cfb9-4bf1-bb53-be9c18606e78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.582650 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-kube-api-access-j8rzd" (OuterVolumeSpecName: "kube-api-access-j8rzd") pod "080e4b8e-cfb9-4bf1-bb53-be9c18606e78" (UID: "080e4b8e-cfb9-4bf1-bb53-be9c18606e78"). InnerVolumeSpecName "kube-api-access-j8rzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.624681 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "080e4b8e-cfb9-4bf1-bb53-be9c18606e78" (UID: "080e4b8e-cfb9-4bf1-bb53-be9c18606e78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.676882 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.676923 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.676941 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8rzd\" (UniqueName: \"kubernetes.io/projected/080e4b8e-cfb9-4bf1-bb53-be9c18606e78-kube-api-access-j8rzd\") on node \"crc\" DevicePath \"\"" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.912166 4741 generic.go:334] "Generic (PLEG): container finished" podID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerID="ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3" exitCode=0 Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.912213 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgvcb" event={"ID":"080e4b8e-cfb9-4bf1-bb53-be9c18606e78","Type":"ContainerDied","Data":"ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3"} Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.912237 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgvcb" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.912265 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgvcb" event={"ID":"080e4b8e-cfb9-4bf1-bb53-be9c18606e78","Type":"ContainerDied","Data":"a1717e68952d297a90ffb1d3aeca8910f5f162da0a58217d6372032bf0ac30c6"} Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.912284 4741 scope.go:117] "RemoveContainer" containerID="ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.950066 4741 scope.go:117] "RemoveContainer" containerID="c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60" Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.950988 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgvcb"] Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.960154 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cgvcb"] Sep 29 21:02:20 crc kubenswrapper[4741]: I0929 21:02:20.988612 4741 scope.go:117] "RemoveContainer" containerID="27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8" Sep 29 21:02:21 crc kubenswrapper[4741]: I0929 21:02:21.021335 4741 scope.go:117] "RemoveContainer" containerID="ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3" Sep 29 21:02:21 crc kubenswrapper[4741]: E0929 21:02:21.021780 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3\": container with ID starting with ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3 not found: ID does not exist" containerID="ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3" Sep 29 21:02:21 crc kubenswrapper[4741]: I0929 21:02:21.021818 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3"} err="failed to get container status \"ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3\": rpc error: code = NotFound desc = could not find container \"ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3\": container with ID starting with ef432459581ede87e40e25a3448879cb3a595ac34570bc39552c1355d04ab6f3 not found: ID does not exist" Sep 29 21:02:21 crc kubenswrapper[4741]: I0929 21:02:21.021845 4741 scope.go:117] "RemoveContainer" containerID="c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60" Sep 29 21:02:21 crc kubenswrapper[4741]: E0929 21:02:21.022143 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60\": container with ID starting with c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60 not found: ID does not exist" containerID="c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60" Sep 29 21:02:21 crc kubenswrapper[4741]: I0929 21:02:21.022171 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60"} err="failed to get container status \"c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60\": rpc error: code = NotFound desc = could not find container \"c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60\": container with ID starting with c700bd60b75790a6aaa9bbdf7e28d4fa7e70db0ae194f9487248b979cda63b60 not found: ID does not exist" Sep 29 21:02:21 crc kubenswrapper[4741]: I0929 21:02:21.022185 4741 scope.go:117] "RemoveContainer" containerID="27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8" Sep 29 21:02:21 crc kubenswrapper[4741]: E0929 21:02:21.022419 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8\": container with ID starting with 27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8 not found: ID does not exist" containerID="27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8" Sep 29 21:02:21 crc kubenswrapper[4741]: I0929 21:02:21.022459 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8"} err="failed to get container status \"27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8\": rpc error: code = NotFound desc = could not find container \"27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8\": container with ID starting with 27c4c53a334b559a5094a414dbcb585e7200ae8eb90be35fdded4a815309e8c8 not found: ID does not exist" Sep 29 21:02:21 crc kubenswrapper[4741]: I0929 21:02:21.103766 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" path="/var/lib/kubelet/pods/080e4b8e-cfb9-4bf1-bb53-be9c18606e78/volumes" Sep 29 21:02:27 crc kubenswrapper[4741]: I0929 21:02:27.032415 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-jtjtg"] Sep 29 21:02:27 crc kubenswrapper[4741]: I0929 21:02:27.043507 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-jtjtg"] Sep 29 21:02:27 crc kubenswrapper[4741]: I0929 21:02:27.100486 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f18429-df86-4df8-aea7-7af91d6e6a2f" path="/var/lib/kubelet/pods/09f18429-df86-4df8-aea7-7af91d6e6a2f/volumes" Sep 29 21:02:28 crc kubenswrapper[4741]: I0929 21:02:28.086438 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:02:28 crc kubenswrapper[4741]: E0929 21:02:28.086652 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:02:31 crc kubenswrapper[4741]: I0929 21:02:31.204611 4741 scope.go:117] "RemoveContainer" containerID="c13ebe7d4bbd4ba511505a2c04b61b885477f90ca7b2510968cc9b0b6d683a66" Sep 29 21:02:31 crc kubenswrapper[4741]: I0929 21:02:31.270353 4741 scope.go:117] "RemoveContainer" containerID="6245bec3c4b5c97c308ec2dd89e7399399b87cb6d9de502362fffb0072c78a7e" Sep 29 21:02:31 crc kubenswrapper[4741]: I0929 21:02:31.312709 4741 scope.go:117] "RemoveContainer" containerID="7c143506f7001f0f1b2cbf17d2e6db522814b3137b760d51e795abc005843ed2" Sep 29 21:02:40 crc kubenswrapper[4741]: I0929 21:02:40.086474 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:02:40 crc kubenswrapper[4741]: E0929 21:02:40.087680 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:02:44 crc kubenswrapper[4741]: I0929 21:02:44.041275 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-lnkrj"] Sep 29 21:02:44 crc kubenswrapper[4741]: I0929 21:02:44.052766 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-lnkrj"] Sep 29 21:02:45 crc kubenswrapper[4741]: I0929 21:02:45.098134 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="353fdc30-16fe-4745-abbd-258072064666" path="/var/lib/kubelet/pods/353fdc30-16fe-4745-abbd-258072064666/volumes" Sep 29 21:02:54 crc kubenswrapper[4741]: I0929 21:02:54.029501 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-4428-account-create-p8bhg"] Sep 29 21:02:54 crc kubenswrapper[4741]: I0929 21:02:54.039623 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-4428-account-create-p8bhg"] Sep 29 21:02:54 crc kubenswrapper[4741]: I0929 21:02:54.086056 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:02:54 crc kubenswrapper[4741]: E0929 21:02:54.086329 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:02:55 crc kubenswrapper[4741]: I0929 21:02:55.100985 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bc712f3-183e-4421-9238-e48591525259" path="/var/lib/kubelet/pods/3bc712f3-183e-4421-9238-e48591525259/volumes" Sep 29 21:03:06 crc kubenswrapper[4741]: I0929 21:03:06.029492 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-tklfv"] Sep 29 21:03:06 crc kubenswrapper[4741]: I0929 21:03:06.037284 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-tklfv"] Sep 29 21:03:07 crc kubenswrapper[4741]: I0929 21:03:07.097315 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a5e675-eeec-4197-b435-fb03bd19662b" path="/var/lib/kubelet/pods/51a5e675-eeec-4197-b435-fb03bd19662b/volumes" Sep 29 21:03:08 crc kubenswrapper[4741]: I0929 21:03:08.085905 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:03:08 crc kubenswrapper[4741]: E0929 21:03:08.086326 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:03:19 crc kubenswrapper[4741]: I0929 21:03:19.096755 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:03:19 crc kubenswrapper[4741]: E0929 21:03:19.097828 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:03:31 crc kubenswrapper[4741]: I0929 21:03:31.087183 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:03:31 crc kubenswrapper[4741]: E0929 21:03:31.088542 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:03:31 crc kubenswrapper[4741]: I0929 21:03:31.439974 4741 scope.go:117] "RemoveContainer" containerID="ee03b83dd6a02bacfbb25bdd8a3ecc5aad2e652eb3d421c96398e4e4f58a3705" Sep 29 21:03:31 crc kubenswrapper[4741]: I0929 21:03:31.470117 4741 scope.go:117] "RemoveContainer" containerID="f66e81ec2e7929dee53f5032ae101d7d8e8a32ab1b65dde91b9aa5d2b9aeffdd" Sep 29 21:03:31 crc kubenswrapper[4741]: I0929 21:03:31.568043 4741 scope.go:117] "RemoveContainer" containerID="8e44993d8028dab55ed0ee5bdec553a980a1503d05a17c10a7b1ec63aff26cb8" Sep 29 21:03:42 crc kubenswrapper[4741]: I0929 21:03:42.086811 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:03:42 crc kubenswrapper[4741]: E0929 21:03:42.087453 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:03:56 crc kubenswrapper[4741]: I0929 21:03:56.086166 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:03:56 crc kubenswrapper[4741]: E0929 21:03:56.086853 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:04:09 crc kubenswrapper[4741]: I0929 21:04:09.094734 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:04:09 crc kubenswrapper[4741]: E0929 21:04:09.095930 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:04:24 crc kubenswrapper[4741]: I0929 21:04:24.086572 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:04:24 crc kubenswrapper[4741]: E0929 21:04:24.087381 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:04:38 crc kubenswrapper[4741]: I0929 21:04:38.085818 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:04:38 crc kubenswrapper[4741]: E0929 21:04:38.086578 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:04:53 crc kubenswrapper[4741]: I0929 21:04:53.085996 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:04:53 crc kubenswrapper[4741]: E0929 21:04:53.086712 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:05:07 crc kubenswrapper[4741]: I0929 21:05:07.085993 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:05:07 crc kubenswrapper[4741]: E0929 21:05:07.086891 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:05:20 crc kubenswrapper[4741]: I0929 21:05:20.085654 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:05:20 crc kubenswrapper[4741]: E0929 21:05:20.086426 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:05:31 crc kubenswrapper[4741]: I0929 21:05:31.086376 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:05:31 crc kubenswrapper[4741]: E0929 21:05:31.087514 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:05:33 crc kubenswrapper[4741]: I0929 21:05:33.934887 4741 generic.go:334] "Generic (PLEG): container finished" podID="e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" containerID="5becc5f6051a5ac17b6b6184af7adee17f732b30c6122a71c44314a685589375" exitCode=0 Sep 29 21:05:33 crc kubenswrapper[4741]: I0929 21:05:33.934972 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" event={"ID":"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261","Type":"ContainerDied","Data":"5becc5f6051a5ac17b6b6184af7adee17f732b30c6122a71c44314a685589375"} Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.361568 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.465853 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-inventory\") pod \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.465944 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksbhb\" (UniqueName: \"kubernetes.io/projected/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-kube-api-access-ksbhb\") pod \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.465968 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ssh-key\") pod \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.466648 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ceph\") pod \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.466795 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-tripleo-cleanup-combined-ca-bundle\") pod \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\" (UID: \"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261\") " Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.471530 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" (UID: "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.471703 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ceph" (OuterVolumeSpecName: "ceph") pod "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" (UID: "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.472153 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-kube-api-access-ksbhb" (OuterVolumeSpecName: "kube-api-access-ksbhb") pod "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" (UID: "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261"). InnerVolumeSpecName "kube-api-access-ksbhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.494720 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" (UID: "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.495426 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-inventory" (OuterVolumeSpecName: "inventory") pod "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" (UID: "e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.569868 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.569900 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksbhb\" (UniqueName: \"kubernetes.io/projected/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-kube-api-access-ksbhb\") on node \"crc\" DevicePath \"\"" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.569910 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.569920 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.569930 4741 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.963366 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" event={"ID":"e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261","Type":"ContainerDied","Data":"4dea27e72a0012f7bd84a84bab136fe7f7995271186636c84e0126872691bf7e"} Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.963435 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5" Sep 29 21:05:35 crc kubenswrapper[4741]: I0929 21:05:35.963446 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dea27e72a0012f7bd84a84bab136fe7f7995271186636c84e0126872691bf7e" Sep 29 21:05:42 crc kubenswrapper[4741]: I0929 21:05:42.087629 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:05:43 crc kubenswrapper[4741]: I0929 21:05:43.046122 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"e6ec2ed5acc4b1e1614d6e88669fc31d76e520de7c8fcff679f157bfd497f398"} Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.021041 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-wfzrx"] Sep 29 21:05:46 crc kubenswrapper[4741]: E0929 21:05:46.023095 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="registry-server" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.023222 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="registry-server" Sep 29 21:05:46 crc kubenswrapper[4741]: E0929 21:05:46.023341 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="extract-content" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.023454 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="extract-content" Sep 29 21:05:46 crc kubenswrapper[4741]: E0929 21:05:46.023553 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.023635 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Sep 29 21:05:46 crc kubenswrapper[4741]: E0929 21:05:46.023756 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="extract-utilities" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.023839 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="extract-utilities" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.024162 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.024260 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="080e4b8e-cfb9-4bf1-bb53-be9c18606e78" containerName="registry-server" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.025142 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.032166 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-wfzrx"] Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.034749 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.034766 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.038736 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.038736 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.093081 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq2bs\" (UniqueName: \"kubernetes.io/projected/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-kube-api-access-rq2bs\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.093950 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.094573 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ceph\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.094903 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.095065 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-inventory\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.196150 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.196250 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ceph\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.196284 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.196316 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-inventory\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.196371 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq2bs\" (UniqueName: \"kubernetes.io/projected/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-kube-api-access-rq2bs\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.206072 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.214958 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ceph\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.215927 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-inventory\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.220306 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.234902 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq2bs\" (UniqueName: \"kubernetes.io/projected/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-kube-api-access-rq2bs\") pod \"bootstrap-openstack-openstack-cell1-wfzrx\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.361295 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.885664 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-wfzrx"] Sep 29 21:05:46 crc kubenswrapper[4741]: W0929 21:05:46.893253 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4fb4b29_ab75_4ef4_be6d_b0c98cf28748.slice/crio-b3e5e16bea2dc0646c8a4efbdf5fc674697a33f5b354257ddc51ff2d15ee0f62 WatchSource:0}: Error finding container b3e5e16bea2dc0646c8a4efbdf5fc674697a33f5b354257ddc51ff2d15ee0f62: Status 404 returned error can't find the container with id b3e5e16bea2dc0646c8a4efbdf5fc674697a33f5b354257ddc51ff2d15ee0f62 Sep 29 21:05:46 crc kubenswrapper[4741]: I0929 21:05:46.898294 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:05:47 crc kubenswrapper[4741]: I0929 21:05:47.109731 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" event={"ID":"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748","Type":"ContainerStarted","Data":"b3e5e16bea2dc0646c8a4efbdf5fc674697a33f5b354257ddc51ff2d15ee0f62"} Sep 29 21:05:48 crc kubenswrapper[4741]: I0929 21:05:48.113197 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" event={"ID":"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748","Type":"ContainerStarted","Data":"81ca1136eb7d58c4823902a2e37883e1a5e2351d8cfe79367341d5a33db1fa04"} Sep 29 21:05:48 crc kubenswrapper[4741]: I0929 21:05:48.145910 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" podStartSLOduration=2.73310122 podStartE2EDuration="3.145888153s" podCreationTimestamp="2025-09-29 21:05:45 +0000 UTC" firstStartedPulling="2025-09-29 21:05:46.898055867 +0000 UTC m=+6988.545845199" lastFinishedPulling="2025-09-29 21:05:47.31084279 +0000 UTC m=+6988.958632132" observedRunningTime="2025-09-29 21:05:48.135650091 +0000 UTC m=+6989.783439463" watchObservedRunningTime="2025-09-29 21:05:48.145888153 +0000 UTC m=+6989.793677495" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.338069 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pwx2j"] Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.341970 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.362537 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwx2j"] Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.439434 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psv5m\" (UniqueName: \"kubernetes.io/projected/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-kube-api-access-psv5m\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.439788 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-catalog-content\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.439896 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-utilities\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.541580 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-catalog-content\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.541620 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-utilities\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.541757 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psv5m\" (UniqueName: \"kubernetes.io/projected/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-kube-api-access-psv5m\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.542166 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-catalog-content\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.542411 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-utilities\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.572807 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psv5m\" (UniqueName: \"kubernetes.io/projected/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-kube-api-access-psv5m\") pod \"redhat-operators-pwx2j\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:30 crc kubenswrapper[4741]: I0929 21:07:30.672609 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:31 crc kubenswrapper[4741]: I0929 21:07:31.143204 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwx2j"] Sep 29 21:07:31 crc kubenswrapper[4741]: I0929 21:07:31.250879 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwx2j" event={"ID":"584fd0a0-6245-45ce-ab60-bcbc0eb43dce","Type":"ContainerStarted","Data":"1c782c6b5abd1a95f2872eb8b61bd3d60c784618ac4ae83fce3728028a85b5bd"} Sep 29 21:07:32 crc kubenswrapper[4741]: I0929 21:07:32.266778 4741 generic.go:334] "Generic (PLEG): container finished" podID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerID="e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4" exitCode=0 Sep 29 21:07:32 crc kubenswrapper[4741]: I0929 21:07:32.266860 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwx2j" event={"ID":"584fd0a0-6245-45ce-ab60-bcbc0eb43dce","Type":"ContainerDied","Data":"e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4"} Sep 29 21:07:34 crc kubenswrapper[4741]: I0929 21:07:34.290215 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwx2j" event={"ID":"584fd0a0-6245-45ce-ab60-bcbc0eb43dce","Type":"ContainerStarted","Data":"7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8"} Sep 29 21:07:37 crc kubenswrapper[4741]: E0929 21:07:37.233948 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod584fd0a0_6245_45ce_ab60_bcbc0eb43dce.slice/crio-conmon-7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8.scope\": RecentStats: unable to find data in memory cache]" Sep 29 21:07:37 crc kubenswrapper[4741]: I0929 21:07:37.317985 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwx2j" event={"ID":"584fd0a0-6245-45ce-ab60-bcbc0eb43dce","Type":"ContainerDied","Data":"7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8"} Sep 29 21:07:37 crc kubenswrapper[4741]: I0929 21:07:37.318004 4741 generic.go:334] "Generic (PLEG): container finished" podID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerID="7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8" exitCode=0 Sep 29 21:07:38 crc kubenswrapper[4741]: I0929 21:07:38.331756 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwx2j" event={"ID":"584fd0a0-6245-45ce-ab60-bcbc0eb43dce","Type":"ContainerStarted","Data":"54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149"} Sep 29 21:07:38 crc kubenswrapper[4741]: I0929 21:07:38.364458 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pwx2j" podStartSLOduration=2.918848503 podStartE2EDuration="8.364439992s" podCreationTimestamp="2025-09-29 21:07:30 +0000 UTC" firstStartedPulling="2025-09-29 21:07:32.270408 +0000 UTC m=+7093.918197332" lastFinishedPulling="2025-09-29 21:07:37.715999489 +0000 UTC m=+7099.363788821" observedRunningTime="2025-09-29 21:07:38.35194559 +0000 UTC m=+7099.999734932" watchObservedRunningTime="2025-09-29 21:07:38.364439992 +0000 UTC m=+7100.012229324" Sep 29 21:07:40 crc kubenswrapper[4741]: I0929 21:07:40.684215 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:40 crc kubenswrapper[4741]: I0929 21:07:40.685647 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:07:41 crc kubenswrapper[4741]: I0929 21:07:41.738523 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pwx2j" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="registry-server" probeResult="failure" output=< Sep 29 21:07:41 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 21:07:41 crc kubenswrapper[4741]: > Sep 29 21:07:51 crc kubenswrapper[4741]: I0929 21:07:51.721158 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pwx2j" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="registry-server" probeResult="failure" output=< Sep 29 21:07:51 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 21:07:51 crc kubenswrapper[4741]: > Sep 29 21:08:01 crc kubenswrapper[4741]: I0929 21:08:01.739757 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:08:01 crc kubenswrapper[4741]: I0929 21:08:01.740411 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:08:01 crc kubenswrapper[4741]: I0929 21:08:01.757686 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pwx2j" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="registry-server" probeResult="failure" output=< Sep 29 21:08:01 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 21:08:01 crc kubenswrapper[4741]: > Sep 29 21:08:10 crc kubenswrapper[4741]: I0929 21:08:10.731184 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:08:10 crc kubenswrapper[4741]: I0929 21:08:10.810607 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:08:11 crc kubenswrapper[4741]: I0929 21:08:11.000878 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwx2j"] Sep 29 21:08:12 crc kubenswrapper[4741]: I0929 21:08:12.697721 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pwx2j" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="registry-server" containerID="cri-o://54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149" gracePeriod=2 Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.215903 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.272858 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-catalog-content\") pod \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.273173 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-utilities\") pod \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.273241 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psv5m\" (UniqueName: \"kubernetes.io/projected/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-kube-api-access-psv5m\") pod \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\" (UID: \"584fd0a0-6245-45ce-ab60-bcbc0eb43dce\") " Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.274274 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-utilities" (OuterVolumeSpecName: "utilities") pod "584fd0a0-6245-45ce-ab60-bcbc0eb43dce" (UID: "584fd0a0-6245-45ce-ab60-bcbc0eb43dce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.275193 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.281464 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-kube-api-access-psv5m" (OuterVolumeSpecName: "kube-api-access-psv5m") pod "584fd0a0-6245-45ce-ab60-bcbc0eb43dce" (UID: "584fd0a0-6245-45ce-ab60-bcbc0eb43dce"). InnerVolumeSpecName "kube-api-access-psv5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.366246 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "584fd0a0-6245-45ce-ab60-bcbc0eb43dce" (UID: "584fd0a0-6245-45ce-ab60-bcbc0eb43dce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.378835 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psv5m\" (UniqueName: \"kubernetes.io/projected/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-kube-api-access-psv5m\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.378860 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584fd0a0-6245-45ce-ab60-bcbc0eb43dce-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.712930 4741 generic.go:334] "Generic (PLEG): container finished" podID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerID="54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149" exitCode=0 Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.712998 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwx2j" event={"ID":"584fd0a0-6245-45ce-ab60-bcbc0eb43dce","Type":"ContainerDied","Data":"54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149"} Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.713014 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwx2j" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.713057 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwx2j" event={"ID":"584fd0a0-6245-45ce-ab60-bcbc0eb43dce","Type":"ContainerDied","Data":"1c782c6b5abd1a95f2872eb8b61bd3d60c784618ac4ae83fce3728028a85b5bd"} Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.713111 4741 scope.go:117] "RemoveContainer" containerID="54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.753118 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwx2j"] Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.764178 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pwx2j"] Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.768984 4741 scope.go:117] "RemoveContainer" containerID="7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.813085 4741 scope.go:117] "RemoveContainer" containerID="e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.869721 4741 scope.go:117] "RemoveContainer" containerID="54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149" Sep 29 21:08:13 crc kubenswrapper[4741]: E0929 21:08:13.870152 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149\": container with ID starting with 54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149 not found: ID does not exist" containerID="54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.870184 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149"} err="failed to get container status \"54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149\": rpc error: code = NotFound desc = could not find container \"54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149\": container with ID starting with 54e044ce2d4605a8276168c5e9b8a9514e190c667e7e748811deba46fdf54149 not found: ID does not exist" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.870226 4741 scope.go:117] "RemoveContainer" containerID="7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8" Sep 29 21:08:13 crc kubenswrapper[4741]: E0929 21:08:13.870850 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8\": container with ID starting with 7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8 not found: ID does not exist" containerID="7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.870890 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8"} err="failed to get container status \"7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8\": rpc error: code = NotFound desc = could not find container \"7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8\": container with ID starting with 7176898581cec0893ecdc7db496ed9c47526ed3e0c14f007faa3722191127ad8 not found: ID does not exist" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.870908 4741 scope.go:117] "RemoveContainer" containerID="e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4" Sep 29 21:08:13 crc kubenswrapper[4741]: E0929 21:08:13.871168 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4\": container with ID starting with e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4 not found: ID does not exist" containerID="e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4" Sep 29 21:08:13 crc kubenswrapper[4741]: I0929 21:08:13.871209 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4"} err="failed to get container status \"e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4\": rpc error: code = NotFound desc = could not find container \"e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4\": container with ID starting with e06b6605d15ad122c36b9f2f07adae92f77896b7979f820a8db2a18e903daec4 not found: ID does not exist" Sep 29 21:08:15 crc kubenswrapper[4741]: I0929 21:08:15.105249 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" path="/var/lib/kubelet/pods/584fd0a0-6245-45ce-ab60-bcbc0eb43dce/volumes" Sep 29 21:08:31 crc kubenswrapper[4741]: I0929 21:08:31.738594 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:08:31 crc kubenswrapper[4741]: I0929 21:08:31.740185 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:08:53 crc kubenswrapper[4741]: I0929 21:08:53.104180 4741 generic.go:334] "Generic (PLEG): container finished" podID="f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" containerID="81ca1136eb7d58c4823902a2e37883e1a5e2351d8cfe79367341d5a33db1fa04" exitCode=0 Sep 29 21:08:53 crc kubenswrapper[4741]: I0929 21:08:53.104238 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" event={"ID":"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748","Type":"ContainerDied","Data":"81ca1136eb7d58c4823902a2e37883e1a5e2351d8cfe79367341d5a33db1fa04"} Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.610155 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.785285 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ceph\") pod \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.785488 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-inventory\") pod \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.785631 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq2bs\" (UniqueName: \"kubernetes.io/projected/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-kube-api-access-rq2bs\") pod \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.785665 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-bootstrap-combined-ca-bundle\") pod \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.785692 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ssh-key\") pod \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\" (UID: \"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748\") " Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.791378 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" (UID: "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.792532 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-kube-api-access-rq2bs" (OuterVolumeSpecName: "kube-api-access-rq2bs") pod "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" (UID: "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748"). InnerVolumeSpecName "kube-api-access-rq2bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.801643 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ceph" (OuterVolumeSpecName: "ceph") pod "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" (UID: "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.820844 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" (UID: "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.835476 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-inventory" (OuterVolumeSpecName: "inventory") pod "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" (UID: "f4fb4b29-ab75-4ef4-be6d-b0c98cf28748"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.887977 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.888018 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.888035 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq2bs\" (UniqueName: \"kubernetes.io/projected/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-kube-api-access-rq2bs\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.888047 4741 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:54 crc kubenswrapper[4741]: I0929 21:08:54.888057 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4fb4b29-ab75-4ef4-be6d-b0c98cf28748-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.122894 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" event={"ID":"f4fb4b29-ab75-4ef4-be6d-b0c98cf28748","Type":"ContainerDied","Data":"b3e5e16bea2dc0646c8a4efbdf5fc674697a33f5b354257ddc51ff2d15ee0f62"} Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.122927 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3e5e16bea2dc0646c8a4efbdf5fc674697a33f5b354257ddc51ff2d15ee0f62" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.122975 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-wfzrx" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.228239 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-g5757"] Sep 29 21:08:55 crc kubenswrapper[4741]: E0929 21:08:55.229074 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="extract-content" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.229093 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="extract-content" Sep 29 21:08:55 crc kubenswrapper[4741]: E0929 21:08:55.229123 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" containerName="bootstrap-openstack-openstack-cell1" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.229132 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" containerName="bootstrap-openstack-openstack-cell1" Sep 29 21:08:55 crc kubenswrapper[4741]: E0929 21:08:55.229146 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="extract-utilities" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.229153 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="extract-utilities" Sep 29 21:08:55 crc kubenswrapper[4741]: E0929 21:08:55.229172 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="registry-server" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.229178 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="registry-server" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.229403 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4fb4b29-ab75-4ef4-be6d-b0c98cf28748" containerName="bootstrap-openstack-openstack-cell1" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.229418 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="584fd0a0-6245-45ce-ab60-bcbc0eb43dce" containerName="registry-server" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.240917 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-g5757"] Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.242150 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.245218 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.245924 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.247016 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.247355 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.396716 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-inventory\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.397202 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ssh-key\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.397365 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ceph\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.397647 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvd2c\" (UniqueName: \"kubernetes.io/projected/80e293dd-446d-4869-babe-815944d08cdb-kube-api-access-nvd2c\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.499618 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvd2c\" (UniqueName: \"kubernetes.io/projected/80e293dd-446d-4869-babe-815944d08cdb-kube-api-access-nvd2c\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.499770 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-inventory\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.499837 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ssh-key\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.499856 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ceph\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.505374 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ceph\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.505672 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ssh-key\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.506361 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-inventory\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.522383 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvd2c\" (UniqueName: \"kubernetes.io/projected/80e293dd-446d-4869-babe-815944d08cdb-kube-api-access-nvd2c\") pod \"download-cache-openstack-openstack-cell1-g5757\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:55 crc kubenswrapper[4741]: I0929 21:08:55.565194 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:08:56 crc kubenswrapper[4741]: I0929 21:08:56.127077 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-g5757"] Sep 29 21:08:56 crc kubenswrapper[4741]: I0929 21:08:56.135598 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-g5757" event={"ID":"80e293dd-446d-4869-babe-815944d08cdb","Type":"ContainerStarted","Data":"6244dac5586b4785c6af866fa0bc34cadd08b998b7abd888406c9f6a7be50370"} Sep 29 21:08:57 crc kubenswrapper[4741]: I0929 21:08:57.149108 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-g5757" event={"ID":"80e293dd-446d-4869-babe-815944d08cdb","Type":"ContainerStarted","Data":"02442ea3df815f1cef8dfded8ef7d0141b72638c4d03a0682c8834e2e1f93b17"} Sep 29 21:08:57 crc kubenswrapper[4741]: I0929 21:08:57.172132 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-g5757" podStartSLOduration=1.580882715 podStartE2EDuration="2.172115902s" podCreationTimestamp="2025-09-29 21:08:55 +0000 UTC" firstStartedPulling="2025-09-29 21:08:56.128064034 +0000 UTC m=+7177.775853366" lastFinishedPulling="2025-09-29 21:08:56.719297221 +0000 UTC m=+7178.367086553" observedRunningTime="2025-09-29 21:08:57.164925605 +0000 UTC m=+7178.812714937" watchObservedRunningTime="2025-09-29 21:08:57.172115902 +0000 UTC m=+7178.819905234" Sep 29 21:09:01 crc kubenswrapper[4741]: I0929 21:09:01.739022 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:09:01 crc kubenswrapper[4741]: I0929 21:09:01.739729 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:09:01 crc kubenswrapper[4741]: I0929 21:09:01.739790 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:09:01 crc kubenswrapper[4741]: I0929 21:09:01.740816 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6ec2ed5acc4b1e1614d6e88669fc31d76e520de7c8fcff679f157bfd497f398"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:09:01 crc kubenswrapper[4741]: I0929 21:09:01.740879 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://e6ec2ed5acc4b1e1614d6e88669fc31d76e520de7c8fcff679f157bfd497f398" gracePeriod=600 Sep 29 21:09:02 crc kubenswrapper[4741]: I0929 21:09:02.208196 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="e6ec2ed5acc4b1e1614d6e88669fc31d76e520de7c8fcff679f157bfd497f398" exitCode=0 Sep 29 21:09:02 crc kubenswrapper[4741]: I0929 21:09:02.208237 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"e6ec2ed5acc4b1e1614d6e88669fc31d76e520de7c8fcff679f157bfd497f398"} Sep 29 21:09:02 crc kubenswrapper[4741]: I0929 21:09:02.208614 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609"} Sep 29 21:09:02 crc kubenswrapper[4741]: I0929 21:09:02.208636 4741 scope.go:117] "RemoveContainer" containerID="cbcf5138a95caf857d35c9443c08af908cd42eb8b227ef77dff3b0d8d8ac8a97" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.185209 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tn74g"] Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.191145 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.200019 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tn74g"] Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.257719 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5frh9\" (UniqueName: \"kubernetes.io/projected/1c4afbb9-eb9b-460d-957e-0c70b3530314-kube-api-access-5frh9\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.258030 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-catalog-content\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.258143 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-utilities\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.360268 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-utilities\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.360444 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5frh9\" (UniqueName: \"kubernetes.io/projected/1c4afbb9-eb9b-460d-957e-0c70b3530314-kube-api-access-5frh9\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.360478 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-catalog-content\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.360812 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-utilities\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.360893 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-catalog-content\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.386343 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5frh9\" (UniqueName: \"kubernetes.io/projected/1c4afbb9-eb9b-460d-957e-0c70b3530314-kube-api-access-5frh9\") pod \"certified-operators-tn74g\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:49 crc kubenswrapper[4741]: I0929 21:09:49.528165 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:50 crc kubenswrapper[4741]: I0929 21:09:50.104636 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tn74g"] Sep 29 21:09:50 crc kubenswrapper[4741]: I0929 21:09:50.718191 4741 generic.go:334] "Generic (PLEG): container finished" podID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerID="66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94" exitCode=0 Sep 29 21:09:50 crc kubenswrapper[4741]: I0929 21:09:50.718252 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn74g" event={"ID":"1c4afbb9-eb9b-460d-957e-0c70b3530314","Type":"ContainerDied","Data":"66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94"} Sep 29 21:09:50 crc kubenswrapper[4741]: I0929 21:09:50.718726 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn74g" event={"ID":"1c4afbb9-eb9b-460d-957e-0c70b3530314","Type":"ContainerStarted","Data":"d7b76fa0b295bcc5f661ebaaa23e1d77eaecac15a584a016853b20eadd663a65"} Sep 29 21:09:52 crc kubenswrapper[4741]: I0929 21:09:52.747552 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn74g" event={"ID":"1c4afbb9-eb9b-460d-957e-0c70b3530314","Type":"ContainerStarted","Data":"6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6"} Sep 29 21:09:53 crc kubenswrapper[4741]: I0929 21:09:53.758690 4741 generic.go:334] "Generic (PLEG): container finished" podID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerID="6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6" exitCode=0 Sep 29 21:09:53 crc kubenswrapper[4741]: I0929 21:09:53.758745 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn74g" event={"ID":"1c4afbb9-eb9b-460d-957e-0c70b3530314","Type":"ContainerDied","Data":"6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6"} Sep 29 21:09:54 crc kubenswrapper[4741]: I0929 21:09:54.770742 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn74g" event={"ID":"1c4afbb9-eb9b-460d-957e-0c70b3530314","Type":"ContainerStarted","Data":"73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be"} Sep 29 21:09:54 crc kubenswrapper[4741]: I0929 21:09:54.800448 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tn74g" podStartSLOduration=2.358873278 podStartE2EDuration="5.800418824s" podCreationTimestamp="2025-09-29 21:09:49 +0000 UTC" firstStartedPulling="2025-09-29 21:09:50.721693408 +0000 UTC m=+7232.369482740" lastFinishedPulling="2025-09-29 21:09:54.163238944 +0000 UTC m=+7235.811028286" observedRunningTime="2025-09-29 21:09:54.795153039 +0000 UTC m=+7236.442942381" watchObservedRunningTime="2025-09-29 21:09:54.800418824 +0000 UTC m=+7236.448208156" Sep 29 21:09:59 crc kubenswrapper[4741]: I0929 21:09:59.528867 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:59 crc kubenswrapper[4741]: I0929 21:09:59.530092 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:59 crc kubenswrapper[4741]: I0929 21:09:59.594917 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:09:59 crc kubenswrapper[4741]: I0929 21:09:59.904140 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.170543 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tn74g"] Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.171714 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tn74g" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="registry-server" containerID="cri-o://73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be" gracePeriod=2 Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.675830 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.762429 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-utilities\") pod \"1c4afbb9-eb9b-460d-957e-0c70b3530314\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.762783 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-catalog-content\") pod \"1c4afbb9-eb9b-460d-957e-0c70b3530314\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.763016 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5frh9\" (UniqueName: \"kubernetes.io/projected/1c4afbb9-eb9b-460d-957e-0c70b3530314-kube-api-access-5frh9\") pod \"1c4afbb9-eb9b-460d-957e-0c70b3530314\" (UID: \"1c4afbb9-eb9b-460d-957e-0c70b3530314\") " Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.763957 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-utilities" (OuterVolumeSpecName: "utilities") pod "1c4afbb9-eb9b-460d-957e-0c70b3530314" (UID: "1c4afbb9-eb9b-460d-957e-0c70b3530314"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.765078 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.768825 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c4afbb9-eb9b-460d-957e-0c70b3530314-kube-api-access-5frh9" (OuterVolumeSpecName: "kube-api-access-5frh9") pod "1c4afbb9-eb9b-460d-957e-0c70b3530314" (UID: "1c4afbb9-eb9b-460d-957e-0c70b3530314"). InnerVolumeSpecName "kube-api-access-5frh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.815737 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c4afbb9-eb9b-460d-957e-0c70b3530314" (UID: "1c4afbb9-eb9b-460d-957e-0c70b3530314"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.867361 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c4afbb9-eb9b-460d-957e-0c70b3530314-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.867432 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5frh9\" (UniqueName: \"kubernetes.io/projected/1c4afbb9-eb9b-460d-957e-0c70b3530314-kube-api-access-5frh9\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.896335 4741 generic.go:334] "Generic (PLEG): container finished" podID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerID="73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be" exitCode=0 Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.896370 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn74g" event={"ID":"1c4afbb9-eb9b-460d-957e-0c70b3530314","Type":"ContainerDied","Data":"73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be"} Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.896408 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn74g" event={"ID":"1c4afbb9-eb9b-460d-957e-0c70b3530314","Type":"ContainerDied","Data":"d7b76fa0b295bcc5f661ebaaa23e1d77eaecac15a584a016853b20eadd663a65"} Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.896443 4741 scope.go:117] "RemoveContainer" containerID="73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.896992 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn74g" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.927143 4741 scope.go:117] "RemoveContainer" containerID="6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.941755 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tn74g"] Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.952145 4741 scope.go:117] "RemoveContainer" containerID="66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.955356 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tn74g"] Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.999373 4741 scope.go:117] "RemoveContainer" containerID="73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be" Sep 29 21:10:02 crc kubenswrapper[4741]: E0929 21:10:02.999865 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be\": container with ID starting with 73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be not found: ID does not exist" containerID="73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.999905 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be"} err="failed to get container status \"73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be\": rpc error: code = NotFound desc = could not find container \"73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be\": container with ID starting with 73b77976d47aba8d3611e9e040bda97e643a0daef43a6050ea8b59128be276be not found: ID does not exist" Sep 29 21:10:02 crc kubenswrapper[4741]: I0929 21:10:02.999931 4741 scope.go:117] "RemoveContainer" containerID="6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6" Sep 29 21:10:03 crc kubenswrapper[4741]: E0929 21:10:03.000441 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6\": container with ID starting with 6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6 not found: ID does not exist" containerID="6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6" Sep 29 21:10:03 crc kubenswrapper[4741]: I0929 21:10:03.000528 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6"} err="failed to get container status \"6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6\": rpc error: code = NotFound desc = could not find container \"6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6\": container with ID starting with 6b4dd5e9e2d0da6dac8503708d3fd2704bcee8c80364c46c02beb8fb16d725d6 not found: ID does not exist" Sep 29 21:10:03 crc kubenswrapper[4741]: I0929 21:10:03.000549 4741 scope.go:117] "RemoveContainer" containerID="66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94" Sep 29 21:10:03 crc kubenswrapper[4741]: E0929 21:10:03.000853 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94\": container with ID starting with 66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94 not found: ID does not exist" containerID="66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94" Sep 29 21:10:03 crc kubenswrapper[4741]: I0929 21:10:03.000884 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94"} err="failed to get container status \"66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94\": rpc error: code = NotFound desc = could not find container \"66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94\": container with ID starting with 66db5ef89ab5917b1c145d0786688ed8ef39be7886b6c8916b8df297cb7abf94 not found: ID does not exist" Sep 29 21:10:03 crc kubenswrapper[4741]: I0929 21:10:03.102790 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" path="/var/lib/kubelet/pods/1c4afbb9-eb9b-460d-957e-0c70b3530314/volumes" Sep 29 21:10:27 crc kubenswrapper[4741]: I0929 21:10:27.938759 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tvn4z"] Sep 29 21:10:27 crc kubenswrapper[4741]: E0929 21:10:27.939905 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="extract-utilities" Sep 29 21:10:27 crc kubenswrapper[4741]: I0929 21:10:27.939920 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="extract-utilities" Sep 29 21:10:27 crc kubenswrapper[4741]: E0929 21:10:27.939932 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="extract-content" Sep 29 21:10:27 crc kubenswrapper[4741]: I0929 21:10:27.939939 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="extract-content" Sep 29 21:10:27 crc kubenswrapper[4741]: E0929 21:10:27.939969 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="registry-server" Sep 29 21:10:27 crc kubenswrapper[4741]: I0929 21:10:27.939980 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="registry-server" Sep 29 21:10:27 crc kubenswrapper[4741]: I0929 21:10:27.940210 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4afbb9-eb9b-460d-957e-0c70b3530314" containerName="registry-server" Sep 29 21:10:27 crc kubenswrapper[4741]: I0929 21:10:27.943646 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:27 crc kubenswrapper[4741]: I0929 21:10:27.954333 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvn4z"] Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.032910 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-catalog-content\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.032961 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bnvb\" (UniqueName: \"kubernetes.io/projected/a80ffdf0-9827-457b-8cdb-de726888c09c-kube-api-access-2bnvb\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.033035 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-utilities\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.135283 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-catalog-content\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.135352 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bnvb\" (UniqueName: \"kubernetes.io/projected/a80ffdf0-9827-457b-8cdb-de726888c09c-kube-api-access-2bnvb\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.135436 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-utilities\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.135804 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-catalog-content\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.135835 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-utilities\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.165487 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bnvb\" (UniqueName: \"kubernetes.io/projected/a80ffdf0-9827-457b-8cdb-de726888c09c-kube-api-access-2bnvb\") pod \"redhat-marketplace-tvn4z\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.262481 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:28 crc kubenswrapper[4741]: I0929 21:10:28.746611 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvn4z"] Sep 29 21:10:29 crc kubenswrapper[4741]: I0929 21:10:29.190564 4741 generic.go:334] "Generic (PLEG): container finished" podID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerID="2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5" exitCode=0 Sep 29 21:10:29 crc kubenswrapper[4741]: I0929 21:10:29.191366 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvn4z" event={"ID":"a80ffdf0-9827-457b-8cdb-de726888c09c","Type":"ContainerDied","Data":"2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5"} Sep 29 21:10:29 crc kubenswrapper[4741]: I0929 21:10:29.191421 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvn4z" event={"ID":"a80ffdf0-9827-457b-8cdb-de726888c09c","Type":"ContainerStarted","Data":"eee41ef18ef7d7babf595717f5d0fe882b9eeb8d9aaa8bede305ca8fdde29dd2"} Sep 29 21:10:31 crc kubenswrapper[4741]: I0929 21:10:31.210018 4741 generic.go:334] "Generic (PLEG): container finished" podID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerID="d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f" exitCode=0 Sep 29 21:10:31 crc kubenswrapper[4741]: I0929 21:10:31.210146 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvn4z" event={"ID":"a80ffdf0-9827-457b-8cdb-de726888c09c","Type":"ContainerDied","Data":"d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f"} Sep 29 21:10:32 crc kubenswrapper[4741]: I0929 21:10:32.227902 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvn4z" event={"ID":"a80ffdf0-9827-457b-8cdb-de726888c09c","Type":"ContainerStarted","Data":"82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576"} Sep 29 21:10:32 crc kubenswrapper[4741]: I0929 21:10:32.231003 4741 generic.go:334] "Generic (PLEG): container finished" podID="80e293dd-446d-4869-babe-815944d08cdb" containerID="02442ea3df815f1cef8dfded8ef7d0141b72638c4d03a0682c8834e2e1f93b17" exitCode=0 Sep 29 21:10:32 crc kubenswrapper[4741]: I0929 21:10:32.231032 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-g5757" event={"ID":"80e293dd-446d-4869-babe-815944d08cdb","Type":"ContainerDied","Data":"02442ea3df815f1cef8dfded8ef7d0141b72638c4d03a0682c8834e2e1f93b17"} Sep 29 21:10:32 crc kubenswrapper[4741]: I0929 21:10:32.265882 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tvn4z" podStartSLOduration=2.72005396 podStartE2EDuration="5.265867087s" podCreationTimestamp="2025-09-29 21:10:27 +0000 UTC" firstStartedPulling="2025-09-29 21:10:29.193319599 +0000 UTC m=+7270.841108951" lastFinishedPulling="2025-09-29 21:10:31.739132746 +0000 UTC m=+7273.386922078" observedRunningTime="2025-09-29 21:10:32.248552823 +0000 UTC m=+7273.896342155" watchObservedRunningTime="2025-09-29 21:10:32.265867087 +0000 UTC m=+7273.913656419" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.770724 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.861020 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvd2c\" (UniqueName: \"kubernetes.io/projected/80e293dd-446d-4869-babe-815944d08cdb-kube-api-access-nvd2c\") pod \"80e293dd-446d-4869-babe-815944d08cdb\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.861259 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ssh-key\") pod \"80e293dd-446d-4869-babe-815944d08cdb\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.861372 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-inventory\") pod \"80e293dd-446d-4869-babe-815944d08cdb\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.861559 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ceph\") pod \"80e293dd-446d-4869-babe-815944d08cdb\" (UID: \"80e293dd-446d-4869-babe-815944d08cdb\") " Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.866230 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e293dd-446d-4869-babe-815944d08cdb-kube-api-access-nvd2c" (OuterVolumeSpecName: "kube-api-access-nvd2c") pod "80e293dd-446d-4869-babe-815944d08cdb" (UID: "80e293dd-446d-4869-babe-815944d08cdb"). InnerVolumeSpecName "kube-api-access-nvd2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.866692 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ceph" (OuterVolumeSpecName: "ceph") pod "80e293dd-446d-4869-babe-815944d08cdb" (UID: "80e293dd-446d-4869-babe-815944d08cdb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.892863 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-inventory" (OuterVolumeSpecName: "inventory") pod "80e293dd-446d-4869-babe-815944d08cdb" (UID: "80e293dd-446d-4869-babe-815944d08cdb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.898585 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80e293dd-446d-4869-babe-815944d08cdb" (UID: "80e293dd-446d-4869-babe-815944d08cdb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.964306 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvd2c\" (UniqueName: \"kubernetes.io/projected/80e293dd-446d-4869-babe-815944d08cdb-kube-api-access-nvd2c\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.964337 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.964346 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:33 crc kubenswrapper[4741]: I0929 21:10:33.964355 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80e293dd-446d-4869-babe-815944d08cdb-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.250689 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-g5757" event={"ID":"80e293dd-446d-4869-babe-815944d08cdb","Type":"ContainerDied","Data":"6244dac5586b4785c6af866fa0bc34cadd08b998b7abd888406c9f6a7be50370"} Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.250728 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6244dac5586b4785c6af866fa0bc34cadd08b998b7abd888406c9f6a7be50370" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.250782 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-g5757" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.356470 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-799wv"] Sep 29 21:10:34 crc kubenswrapper[4741]: E0929 21:10:34.357324 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e293dd-446d-4869-babe-815944d08cdb" containerName="download-cache-openstack-openstack-cell1" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.357347 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e293dd-446d-4869-babe-815944d08cdb" containerName="download-cache-openstack-openstack-cell1" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.357714 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e293dd-446d-4869-babe-815944d08cdb" containerName="download-cache-openstack-openstack-cell1" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.358874 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.362648 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.362894 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.362978 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.362985 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.367053 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-799wv"] Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.370945 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ceph\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.371013 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zq8l\" (UniqueName: \"kubernetes.io/projected/437a18a3-85a4-4c95-87cd-bfc208f23a23-kube-api-access-8zq8l\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.371041 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ssh-key\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.371211 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-inventory\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.472381 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ceph\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.472477 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zq8l\" (UniqueName: \"kubernetes.io/projected/437a18a3-85a4-4c95-87cd-bfc208f23a23-kube-api-access-8zq8l\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.472502 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ssh-key\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.472572 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-inventory\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.477917 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ceph\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.479058 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ssh-key\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.479440 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-inventory\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.489771 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zq8l\" (UniqueName: \"kubernetes.io/projected/437a18a3-85a4-4c95-87cd-bfc208f23a23-kube-api-access-8zq8l\") pod \"configure-network-openstack-openstack-cell1-799wv\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:34 crc kubenswrapper[4741]: I0929 21:10:34.689316 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:10:35 crc kubenswrapper[4741]: I0929 21:10:35.258455 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-799wv"] Sep 29 21:10:36 crc kubenswrapper[4741]: I0929 21:10:36.273926 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-799wv" event={"ID":"437a18a3-85a4-4c95-87cd-bfc208f23a23","Type":"ContainerStarted","Data":"30bfa6f684ba417c2107107618ce5234ec0835a29c47b4ad8af7876874e7f2f5"} Sep 29 21:10:38 crc kubenswrapper[4741]: I0929 21:10:38.263523 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:38 crc kubenswrapper[4741]: I0929 21:10:38.263865 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:38 crc kubenswrapper[4741]: I0929 21:10:38.308480 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:39 crc kubenswrapper[4741]: I0929 21:10:39.352464 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:39 crc kubenswrapper[4741]: I0929 21:10:39.402470 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvn4z"] Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.326997 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tvn4z" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="registry-server" containerID="cri-o://82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576" gracePeriod=2 Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.868696 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.885289 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-catalog-content\") pod \"a80ffdf0-9827-457b-8cdb-de726888c09c\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.885437 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-utilities\") pod \"a80ffdf0-9827-457b-8cdb-de726888c09c\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.885767 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bnvb\" (UniqueName: \"kubernetes.io/projected/a80ffdf0-9827-457b-8cdb-de726888c09c-kube-api-access-2bnvb\") pod \"a80ffdf0-9827-457b-8cdb-de726888c09c\" (UID: \"a80ffdf0-9827-457b-8cdb-de726888c09c\") " Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.886376 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-utilities" (OuterVolumeSpecName: "utilities") pod "a80ffdf0-9827-457b-8cdb-de726888c09c" (UID: "a80ffdf0-9827-457b-8cdb-de726888c09c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.886858 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.892046 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a80ffdf0-9827-457b-8cdb-de726888c09c-kube-api-access-2bnvb" (OuterVolumeSpecName: "kube-api-access-2bnvb") pod "a80ffdf0-9827-457b-8cdb-de726888c09c" (UID: "a80ffdf0-9827-457b-8cdb-de726888c09c"). InnerVolumeSpecName "kube-api-access-2bnvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.915472 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a80ffdf0-9827-457b-8cdb-de726888c09c" (UID: "a80ffdf0-9827-457b-8cdb-de726888c09c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.988272 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bnvb\" (UniqueName: \"kubernetes.io/projected/a80ffdf0-9827-457b-8cdb-de726888c09c-kube-api-access-2bnvb\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:41 crc kubenswrapper[4741]: I0929 21:10:41.988317 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80ffdf0-9827-457b-8cdb-de726888c09c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.342689 4741 generic.go:334] "Generic (PLEG): container finished" podID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerID="82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576" exitCode=0 Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.342763 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvn4z" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.342786 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvn4z" event={"ID":"a80ffdf0-9827-457b-8cdb-de726888c09c","Type":"ContainerDied","Data":"82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576"} Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.343118 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvn4z" event={"ID":"a80ffdf0-9827-457b-8cdb-de726888c09c","Type":"ContainerDied","Data":"eee41ef18ef7d7babf595717f5d0fe882b9eeb8d9aaa8bede305ca8fdde29dd2"} Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.343164 4741 scope.go:117] "RemoveContainer" containerID="82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.366951 4741 scope.go:117] "RemoveContainer" containerID="d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.380039 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvn4z"] Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.389369 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvn4z"] Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.408329 4741 scope.go:117] "RemoveContainer" containerID="2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.437547 4741 scope.go:117] "RemoveContainer" containerID="82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576" Sep 29 21:10:42 crc kubenswrapper[4741]: E0929 21:10:42.437926 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576\": container with ID starting with 82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576 not found: ID does not exist" containerID="82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.437961 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576"} err="failed to get container status \"82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576\": rpc error: code = NotFound desc = could not find container \"82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576\": container with ID starting with 82c9ead849ca99e91d982c29cf5be41050841c7eb6183609441212e55fb2a576 not found: ID does not exist" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.437982 4741 scope.go:117] "RemoveContainer" containerID="d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f" Sep 29 21:10:42 crc kubenswrapper[4741]: E0929 21:10:42.438225 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f\": container with ID starting with d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f not found: ID does not exist" containerID="d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.438246 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f"} err="failed to get container status \"d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f\": rpc error: code = NotFound desc = could not find container \"d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f\": container with ID starting with d360ddc7794ca7a985998450a400d569c5eb5d8c1f4a1df78d9ed8ae90e3fb2f not found: ID does not exist" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.438259 4741 scope.go:117] "RemoveContainer" containerID="2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5" Sep 29 21:10:42 crc kubenswrapper[4741]: E0929 21:10:42.438491 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5\": container with ID starting with 2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5 not found: ID does not exist" containerID="2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5" Sep 29 21:10:42 crc kubenswrapper[4741]: I0929 21:10:42.438514 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5"} err="failed to get container status \"2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5\": rpc error: code = NotFound desc = could not find container \"2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5\": container with ID starting with 2afb7d20e566de77b80c6ae8f453e9d6d2f41bb31857821f1eee5bfaa2ce1af5 not found: ID does not exist" Sep 29 21:10:43 crc kubenswrapper[4741]: I0929 21:10:43.106085 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" path="/var/lib/kubelet/pods/a80ffdf0-9827-457b-8cdb-de726888c09c/volumes" Sep 29 21:10:49 crc kubenswrapper[4741]: I0929 21:10:49.417784 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-799wv" event={"ID":"437a18a3-85a4-4c95-87cd-bfc208f23a23","Type":"ContainerStarted","Data":"8f67e2ee2159b8300353ee717b8b4caefeba9202a73de906efaab7d33f97829e"} Sep 29 21:10:49 crc kubenswrapper[4741]: I0929 21:10:49.442992 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-799wv" podStartSLOduration=1.8458462199999999 podStartE2EDuration="15.442972896s" podCreationTimestamp="2025-09-29 21:10:34 +0000 UTC" firstStartedPulling="2025-09-29 21:10:35.286738152 +0000 UTC m=+7276.934527484" lastFinishedPulling="2025-09-29 21:10:48.883864838 +0000 UTC m=+7290.531654160" observedRunningTime="2025-09-29 21:10:49.435670706 +0000 UTC m=+7291.083460038" watchObservedRunningTime="2025-09-29 21:10:49.442972896 +0000 UTC m=+7291.090762228" Sep 29 21:11:31 crc kubenswrapper[4741]: I0929 21:11:31.739500 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:11:31 crc kubenswrapper[4741]: I0929 21:11:31.739993 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:12:01 crc kubenswrapper[4741]: I0929 21:12:01.741007 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:12:01 crc kubenswrapper[4741]: I0929 21:12:01.741635 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:12:08 crc kubenswrapper[4741]: I0929 21:12:08.289580 4741 generic.go:334] "Generic (PLEG): container finished" podID="437a18a3-85a4-4c95-87cd-bfc208f23a23" containerID="8f67e2ee2159b8300353ee717b8b4caefeba9202a73de906efaab7d33f97829e" exitCode=0 Sep 29 21:12:08 crc kubenswrapper[4741]: I0929 21:12:08.289628 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-799wv" event={"ID":"437a18a3-85a4-4c95-87cd-bfc208f23a23","Type":"ContainerDied","Data":"8f67e2ee2159b8300353ee717b8b4caefeba9202a73de906efaab7d33f97829e"} Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.857336 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.925292 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-inventory\") pod \"437a18a3-85a4-4c95-87cd-bfc208f23a23\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.925370 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ceph\") pod \"437a18a3-85a4-4c95-87cd-bfc208f23a23\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.925420 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zq8l\" (UniqueName: \"kubernetes.io/projected/437a18a3-85a4-4c95-87cd-bfc208f23a23-kube-api-access-8zq8l\") pod \"437a18a3-85a4-4c95-87cd-bfc208f23a23\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.925441 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ssh-key\") pod \"437a18a3-85a4-4c95-87cd-bfc208f23a23\" (UID: \"437a18a3-85a4-4c95-87cd-bfc208f23a23\") " Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.932083 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ceph" (OuterVolumeSpecName: "ceph") pod "437a18a3-85a4-4c95-87cd-bfc208f23a23" (UID: "437a18a3-85a4-4c95-87cd-bfc208f23a23"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.949250 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/437a18a3-85a4-4c95-87cd-bfc208f23a23-kube-api-access-8zq8l" (OuterVolumeSpecName: "kube-api-access-8zq8l") pod "437a18a3-85a4-4c95-87cd-bfc208f23a23" (UID: "437a18a3-85a4-4c95-87cd-bfc208f23a23"). InnerVolumeSpecName "kube-api-access-8zq8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.957693 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-inventory" (OuterVolumeSpecName: "inventory") pod "437a18a3-85a4-4c95-87cd-bfc208f23a23" (UID: "437a18a3-85a4-4c95-87cd-bfc208f23a23"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:12:09 crc kubenswrapper[4741]: I0929 21:12:09.962435 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "437a18a3-85a4-4c95-87cd-bfc208f23a23" (UID: "437a18a3-85a4-4c95-87cd-bfc208f23a23"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.027517 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.027557 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.027568 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zq8l\" (UniqueName: \"kubernetes.io/projected/437a18a3-85a4-4c95-87cd-bfc208f23a23-kube-api-access-8zq8l\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.027581 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/437a18a3-85a4-4c95-87cd-bfc208f23a23-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.330874 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-799wv" event={"ID":"437a18a3-85a4-4c95-87cd-bfc208f23a23","Type":"ContainerDied","Data":"30bfa6f684ba417c2107107618ce5234ec0835a29c47b4ad8af7876874e7f2f5"} Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.331291 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30bfa6f684ba417c2107107618ce5234ec0835a29c47b4ad8af7876874e7f2f5" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.330923 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-799wv" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.411325 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-wqv85"] Sep 29 21:12:10 crc kubenswrapper[4741]: E0929 21:12:10.411843 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="registry-server" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.411864 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="registry-server" Sep 29 21:12:10 crc kubenswrapper[4741]: E0929 21:12:10.411900 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="extract-utilities" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.411908 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="extract-utilities" Sep 29 21:12:10 crc kubenswrapper[4741]: E0929 21:12:10.411925 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="extract-content" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.411932 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="extract-content" Sep 29 21:12:10 crc kubenswrapper[4741]: E0929 21:12:10.411950 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437a18a3-85a4-4c95-87cd-bfc208f23a23" containerName="configure-network-openstack-openstack-cell1" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.411957 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="437a18a3-85a4-4c95-87cd-bfc208f23a23" containerName="configure-network-openstack-openstack-cell1" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.412173 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="437a18a3-85a4-4c95-87cd-bfc208f23a23" containerName="configure-network-openstack-openstack-cell1" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.412196 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a80ffdf0-9827-457b-8cdb-de726888c09c" containerName="registry-server" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.413097 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.416697 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.416823 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.416831 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.416874 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.428195 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-wqv85"] Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.536159 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ssh-key\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.536334 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ceph\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.536405 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-inventory\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.536454 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4tck\" (UniqueName: \"kubernetes.io/projected/b531de97-929c-4a21-af0b-fa956d7b6cc6-kube-api-access-l4tck\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.638383 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ssh-key\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.638557 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ceph\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.639363 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-inventory\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.639443 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4tck\" (UniqueName: \"kubernetes.io/projected/b531de97-929c-4a21-af0b-fa956d7b6cc6-kube-api-access-l4tck\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.644077 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-inventory\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.644075 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ssh-key\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.653687 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ceph\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.657773 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4tck\" (UniqueName: \"kubernetes.io/projected/b531de97-929c-4a21-af0b-fa956d7b6cc6-kube-api-access-l4tck\") pod \"validate-network-openstack-openstack-cell1-wqv85\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:10 crc kubenswrapper[4741]: I0929 21:12:10.733445 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:11 crc kubenswrapper[4741]: I0929 21:12:11.275670 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-wqv85"] Sep 29 21:12:11 crc kubenswrapper[4741]: W0929 21:12:11.279683 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb531de97_929c_4a21_af0b_fa956d7b6cc6.slice/crio-0d8dbadd088fb3f1aed807bfabf30fc2de08dba211c7104bc6e829e40a74d073 WatchSource:0}: Error finding container 0d8dbadd088fb3f1aed807bfabf30fc2de08dba211c7104bc6e829e40a74d073: Status 404 returned error can't find the container with id 0d8dbadd088fb3f1aed807bfabf30fc2de08dba211c7104bc6e829e40a74d073 Sep 29 21:12:11 crc kubenswrapper[4741]: I0929 21:12:11.282920 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:12:11 crc kubenswrapper[4741]: I0929 21:12:11.345037 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" event={"ID":"b531de97-929c-4a21-af0b-fa956d7b6cc6","Type":"ContainerStarted","Data":"0d8dbadd088fb3f1aed807bfabf30fc2de08dba211c7104bc6e829e40a74d073"} Sep 29 21:12:12 crc kubenswrapper[4741]: I0929 21:12:12.357811 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" event={"ID":"b531de97-929c-4a21-af0b-fa956d7b6cc6","Type":"ContainerStarted","Data":"9890a4de59d938d2c4f0f7e6784d357245d28215d1a8deb2c54345986907bf0c"} Sep 29 21:12:12 crc kubenswrapper[4741]: I0929 21:12:12.393438 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" podStartSLOduration=1.91679215 podStartE2EDuration="2.393416957s" podCreationTimestamp="2025-09-29 21:12:10 +0000 UTC" firstStartedPulling="2025-09-29 21:12:11.282564213 +0000 UTC m=+7372.930353545" lastFinishedPulling="2025-09-29 21:12:11.75918902 +0000 UTC m=+7373.406978352" observedRunningTime="2025-09-29 21:12:12.377962632 +0000 UTC m=+7374.025751994" watchObservedRunningTime="2025-09-29 21:12:12.393416957 +0000 UTC m=+7374.041206289" Sep 29 21:12:17 crc kubenswrapper[4741]: I0929 21:12:17.413920 4741 generic.go:334] "Generic (PLEG): container finished" podID="b531de97-929c-4a21-af0b-fa956d7b6cc6" containerID="9890a4de59d938d2c4f0f7e6784d357245d28215d1a8deb2c54345986907bf0c" exitCode=0 Sep 29 21:12:17 crc kubenswrapper[4741]: I0929 21:12:17.414028 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" event={"ID":"b531de97-929c-4a21-af0b-fa956d7b6cc6","Type":"ContainerDied","Data":"9890a4de59d938d2c4f0f7e6784d357245d28215d1a8deb2c54345986907bf0c"} Sep 29 21:12:18 crc kubenswrapper[4741]: I0929 21:12:18.866669 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.029523 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ceph\") pod \"b531de97-929c-4a21-af0b-fa956d7b6cc6\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.029997 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-inventory\") pod \"b531de97-929c-4a21-af0b-fa956d7b6cc6\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.030045 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4tck\" (UniqueName: \"kubernetes.io/projected/b531de97-929c-4a21-af0b-fa956d7b6cc6-kube-api-access-l4tck\") pod \"b531de97-929c-4a21-af0b-fa956d7b6cc6\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.030109 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ssh-key\") pod \"b531de97-929c-4a21-af0b-fa956d7b6cc6\" (UID: \"b531de97-929c-4a21-af0b-fa956d7b6cc6\") " Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.035406 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ceph" (OuterVolumeSpecName: "ceph") pod "b531de97-929c-4a21-af0b-fa956d7b6cc6" (UID: "b531de97-929c-4a21-af0b-fa956d7b6cc6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.036370 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b531de97-929c-4a21-af0b-fa956d7b6cc6-kube-api-access-l4tck" (OuterVolumeSpecName: "kube-api-access-l4tck") pod "b531de97-929c-4a21-af0b-fa956d7b6cc6" (UID: "b531de97-929c-4a21-af0b-fa956d7b6cc6"). InnerVolumeSpecName "kube-api-access-l4tck". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.062073 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-inventory" (OuterVolumeSpecName: "inventory") pod "b531de97-929c-4a21-af0b-fa956d7b6cc6" (UID: "b531de97-929c-4a21-af0b-fa956d7b6cc6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.091663 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b531de97-929c-4a21-af0b-fa956d7b6cc6" (UID: "b531de97-929c-4a21-af0b-fa956d7b6cc6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.133456 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.133491 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4tck\" (UniqueName: \"kubernetes.io/projected/b531de97-929c-4a21-af0b-fa956d7b6cc6-kube-api-access-l4tck\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.133505 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.133547 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b531de97-929c-4a21-af0b-fa956d7b6cc6-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.437159 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" event={"ID":"b531de97-929c-4a21-af0b-fa956d7b6cc6","Type":"ContainerDied","Data":"0d8dbadd088fb3f1aed807bfabf30fc2de08dba211c7104bc6e829e40a74d073"} Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.437203 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d8dbadd088fb3f1aed807bfabf30fc2de08dba211c7104bc6e829e40a74d073" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.437240 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-wqv85" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.497014 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-krgtk"] Sep 29 21:12:19 crc kubenswrapper[4741]: E0929 21:12:19.497552 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b531de97-929c-4a21-af0b-fa956d7b6cc6" containerName="validate-network-openstack-openstack-cell1" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.497572 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b531de97-929c-4a21-af0b-fa956d7b6cc6" containerName="validate-network-openstack-openstack-cell1" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.497838 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b531de97-929c-4a21-af0b-fa956d7b6cc6" containerName="validate-network-openstack-openstack-cell1" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.498890 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.501137 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.501377 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.503097 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.503636 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.509156 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-krgtk"] Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.642973 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ceph\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.643689 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ssh-key\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.644114 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-inventory\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.644214 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvmhv\" (UniqueName: \"kubernetes.io/projected/eac2989f-b47f-4ded-9a54-b8abccfb14bb-kube-api-access-gvmhv\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.747798 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-inventory\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.747990 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvmhv\" (UniqueName: \"kubernetes.io/projected/eac2989f-b47f-4ded-9a54-b8abccfb14bb-kube-api-access-gvmhv\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.748137 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ceph\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.748292 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ssh-key\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.751439 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-inventory\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.751885 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ceph\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.752172 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ssh-key\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.765508 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvmhv\" (UniqueName: \"kubernetes.io/projected/eac2989f-b47f-4ded-9a54-b8abccfb14bb-kube-api-access-gvmhv\") pod \"install-os-openstack-openstack-cell1-krgtk\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:19 crc kubenswrapper[4741]: I0929 21:12:19.828088 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:12:20 crc kubenswrapper[4741]: I0929 21:12:20.386406 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-krgtk"] Sep 29 21:12:20 crc kubenswrapper[4741]: W0929 21:12:20.388591 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeac2989f_b47f_4ded_9a54_b8abccfb14bb.slice/crio-9e298f30580e1048a12caa9d35c14c0b6fd4eccf7b0f6680e0ce6899db8a65f1 WatchSource:0}: Error finding container 9e298f30580e1048a12caa9d35c14c0b6fd4eccf7b0f6680e0ce6899db8a65f1: Status 404 returned error can't find the container with id 9e298f30580e1048a12caa9d35c14c0b6fd4eccf7b0f6680e0ce6899db8a65f1 Sep 29 21:12:20 crc kubenswrapper[4741]: I0929 21:12:20.447545 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-krgtk" event={"ID":"eac2989f-b47f-4ded-9a54-b8abccfb14bb","Type":"ContainerStarted","Data":"9e298f30580e1048a12caa9d35c14c0b6fd4eccf7b0f6680e0ce6899db8a65f1"} Sep 29 21:12:21 crc kubenswrapper[4741]: I0929 21:12:21.456731 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-krgtk" event={"ID":"eac2989f-b47f-4ded-9a54-b8abccfb14bb","Type":"ContainerStarted","Data":"03abb0461e0aeda3bc950603f6dfc16c8779d5aba699ccbe9cf6950545a5be12"} Sep 29 21:12:31 crc kubenswrapper[4741]: I0929 21:12:31.739253 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:12:31 crc kubenswrapper[4741]: I0929 21:12:31.739834 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:12:31 crc kubenswrapper[4741]: I0929 21:12:31.739955 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:12:31 crc kubenswrapper[4741]: I0929 21:12:31.740893 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:12:31 crc kubenswrapper[4741]: I0929 21:12:31.741050 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" gracePeriod=600 Sep 29 21:12:31 crc kubenswrapper[4741]: E0929 21:12:31.862856 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:12:32 crc kubenswrapper[4741]: I0929 21:12:32.561046 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" exitCode=0 Sep 29 21:12:32 crc kubenswrapper[4741]: I0929 21:12:32.561092 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609"} Sep 29 21:12:32 crc kubenswrapper[4741]: I0929 21:12:32.561133 4741 scope.go:117] "RemoveContainer" containerID="e6ec2ed5acc4b1e1614d6e88669fc31d76e520de7c8fcff679f157bfd497f398" Sep 29 21:12:32 crc kubenswrapper[4741]: I0929 21:12:32.561971 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:12:32 crc kubenswrapper[4741]: E0929 21:12:32.562652 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:12:32 crc kubenswrapper[4741]: I0929 21:12:32.583943 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-krgtk" podStartSLOduration=12.901855917 podStartE2EDuration="13.583922327s" podCreationTimestamp="2025-09-29 21:12:19 +0000 UTC" firstStartedPulling="2025-09-29 21:12:20.390463884 +0000 UTC m=+7382.038253216" lastFinishedPulling="2025-09-29 21:12:21.072530294 +0000 UTC m=+7382.720319626" observedRunningTime="2025-09-29 21:12:21.474768613 +0000 UTC m=+7383.122557945" watchObservedRunningTime="2025-09-29 21:12:32.583922327 +0000 UTC m=+7394.231711659" Sep 29 21:12:48 crc kubenswrapper[4741]: I0929 21:12:48.086022 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:12:48 crc kubenswrapper[4741]: E0929 21:12:48.087019 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:13:00 crc kubenswrapper[4741]: I0929 21:13:00.085566 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:13:00 crc kubenswrapper[4741]: E0929 21:13:00.086338 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:13:03 crc kubenswrapper[4741]: I0929 21:13:03.883607 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wbqfj"] Sep 29 21:13:03 crc kubenswrapper[4741]: I0929 21:13:03.890643 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:03 crc kubenswrapper[4741]: I0929 21:13:03.907368 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbqfj"] Sep 29 21:13:03 crc kubenswrapper[4741]: I0929 21:13:03.963890 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbjmr\" (UniqueName: \"kubernetes.io/projected/aa4eb3c5-e143-4aec-8cd6-6446638f4900-kube-api-access-mbjmr\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:03 crc kubenswrapper[4741]: I0929 21:13:03.964239 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-catalog-content\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:03 crc kubenswrapper[4741]: I0929 21:13:03.964529 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-utilities\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.067080 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-catalog-content\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.067215 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-utilities\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.067338 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbjmr\" (UniqueName: \"kubernetes.io/projected/aa4eb3c5-e143-4aec-8cd6-6446638f4900-kube-api-access-mbjmr\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.068001 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-utilities\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.068174 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-catalog-content\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.112193 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbjmr\" (UniqueName: \"kubernetes.io/projected/aa4eb3c5-e143-4aec-8cd6-6446638f4900-kube-api-access-mbjmr\") pod \"community-operators-wbqfj\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.225075 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.770569 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbqfj"] Sep 29 21:13:04 crc kubenswrapper[4741]: I0929 21:13:04.928193 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqfj" event={"ID":"aa4eb3c5-e143-4aec-8cd6-6446638f4900","Type":"ContainerStarted","Data":"67db57a858a054b555e9466b0233c45bc7d8cbf62dd3a00ac6f0862cf61a239b"} Sep 29 21:13:05 crc kubenswrapper[4741]: I0929 21:13:05.938924 4741 generic.go:334] "Generic (PLEG): container finished" podID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerID="b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1" exitCode=0 Sep 29 21:13:05 crc kubenswrapper[4741]: I0929 21:13:05.938989 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqfj" event={"ID":"aa4eb3c5-e143-4aec-8cd6-6446638f4900","Type":"ContainerDied","Data":"b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1"} Sep 29 21:13:07 crc kubenswrapper[4741]: I0929 21:13:07.983594 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqfj" event={"ID":"aa4eb3c5-e143-4aec-8cd6-6446638f4900","Type":"ContainerStarted","Data":"669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46"} Sep 29 21:13:09 crc kubenswrapper[4741]: I0929 21:13:09.009029 4741 generic.go:334] "Generic (PLEG): container finished" podID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerID="669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46" exitCode=0 Sep 29 21:13:09 crc kubenswrapper[4741]: I0929 21:13:09.009086 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqfj" event={"ID":"aa4eb3c5-e143-4aec-8cd6-6446638f4900","Type":"ContainerDied","Data":"669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46"} Sep 29 21:13:10 crc kubenswrapper[4741]: I0929 21:13:10.021229 4741 generic.go:334] "Generic (PLEG): container finished" podID="eac2989f-b47f-4ded-9a54-b8abccfb14bb" containerID="03abb0461e0aeda3bc950603f6dfc16c8779d5aba699ccbe9cf6950545a5be12" exitCode=0 Sep 29 21:13:10 crc kubenswrapper[4741]: I0929 21:13:10.021321 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-krgtk" event={"ID":"eac2989f-b47f-4ded-9a54-b8abccfb14bb","Type":"ContainerDied","Data":"03abb0461e0aeda3bc950603f6dfc16c8779d5aba699ccbe9cf6950545a5be12"} Sep 29 21:13:10 crc kubenswrapper[4741]: I0929 21:13:10.025369 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqfj" event={"ID":"aa4eb3c5-e143-4aec-8cd6-6446638f4900","Type":"ContainerStarted","Data":"b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7"} Sep 29 21:13:10 crc kubenswrapper[4741]: I0929 21:13:10.069773 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wbqfj" podStartSLOduration=3.5801164610000002 podStartE2EDuration="7.069753695s" podCreationTimestamp="2025-09-29 21:13:03 +0000 UTC" firstStartedPulling="2025-09-29 21:13:05.94172403 +0000 UTC m=+7427.589513362" lastFinishedPulling="2025-09-29 21:13:09.431361264 +0000 UTC m=+7431.079150596" observedRunningTime="2025-09-29 21:13:10.062844739 +0000 UTC m=+7431.710634071" watchObservedRunningTime="2025-09-29 21:13:10.069753695 +0000 UTC m=+7431.717543037" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.586979 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.658016 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ssh-key\") pod \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.658273 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ceph\") pod \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.658343 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvmhv\" (UniqueName: \"kubernetes.io/projected/eac2989f-b47f-4ded-9a54-b8abccfb14bb-kube-api-access-gvmhv\") pod \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.658409 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-inventory\") pod \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\" (UID: \"eac2989f-b47f-4ded-9a54-b8abccfb14bb\") " Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.664079 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac2989f-b47f-4ded-9a54-b8abccfb14bb-kube-api-access-gvmhv" (OuterVolumeSpecName: "kube-api-access-gvmhv") pod "eac2989f-b47f-4ded-9a54-b8abccfb14bb" (UID: "eac2989f-b47f-4ded-9a54-b8abccfb14bb"). InnerVolumeSpecName "kube-api-access-gvmhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.664329 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ceph" (OuterVolumeSpecName: "ceph") pod "eac2989f-b47f-4ded-9a54-b8abccfb14bb" (UID: "eac2989f-b47f-4ded-9a54-b8abccfb14bb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.694604 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-inventory" (OuterVolumeSpecName: "inventory") pod "eac2989f-b47f-4ded-9a54-b8abccfb14bb" (UID: "eac2989f-b47f-4ded-9a54-b8abccfb14bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.713264 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eac2989f-b47f-4ded-9a54-b8abccfb14bb" (UID: "eac2989f-b47f-4ded-9a54-b8abccfb14bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.761484 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.761534 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.761555 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvmhv\" (UniqueName: \"kubernetes.io/projected/eac2989f-b47f-4ded-9a54-b8abccfb14bb-kube-api-access-gvmhv\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:11 crc kubenswrapper[4741]: I0929 21:13:11.761576 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac2989f-b47f-4ded-9a54-b8abccfb14bb-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.051624 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-krgtk" event={"ID":"eac2989f-b47f-4ded-9a54-b8abccfb14bb","Type":"ContainerDied","Data":"9e298f30580e1048a12caa9d35c14c0b6fd4eccf7b0f6680e0ce6899db8a65f1"} Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.051665 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-krgtk" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.051684 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e298f30580e1048a12caa9d35c14c0b6fd4eccf7b0f6680e0ce6899db8a65f1" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.087635 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:13:12 crc kubenswrapper[4741]: E0929 21:13:12.088272 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.163473 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hwlcx"] Sep 29 21:13:12 crc kubenswrapper[4741]: E0929 21:13:12.164270 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac2989f-b47f-4ded-9a54-b8abccfb14bb" containerName="install-os-openstack-openstack-cell1" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.164306 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac2989f-b47f-4ded-9a54-b8abccfb14bb" containerName="install-os-openstack-openstack-cell1" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.164716 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac2989f-b47f-4ded-9a54-b8abccfb14bb" containerName="install-os-openstack-openstack-cell1" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.166145 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.170034 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.170117 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.170037 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.170331 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.175621 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hwlcx"] Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.273838 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ceph\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.274006 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.274172 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw9nv\" (UniqueName: \"kubernetes.io/projected/f372c5ae-6094-4c36-8b23-0a06a4e50c93-kube-api-access-xw9nv\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.274271 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-inventory\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.376807 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw9nv\" (UniqueName: \"kubernetes.io/projected/f372c5ae-6094-4c36-8b23-0a06a4e50c93-kube-api-access-xw9nv\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.376893 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-inventory\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.376945 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ceph\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.377020 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.382668 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-inventory\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.383252 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.385330 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ceph\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.402178 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw9nv\" (UniqueName: \"kubernetes.io/projected/f372c5ae-6094-4c36-8b23-0a06a4e50c93-kube-api-access-xw9nv\") pod \"configure-os-openstack-openstack-cell1-hwlcx\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:12 crc kubenswrapper[4741]: I0929 21:13:12.490958 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:13 crc kubenswrapper[4741]: I0929 21:13:13.077738 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hwlcx"] Sep 29 21:13:13 crc kubenswrapper[4741]: W0929 21:13:13.087566 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf372c5ae_6094_4c36_8b23_0a06a4e50c93.slice/crio-c2ad74d32f8657dd09a746af4d845661390188b07f6a65c51c1ef06af45dd98a WatchSource:0}: Error finding container c2ad74d32f8657dd09a746af4d845661390188b07f6a65c51c1ef06af45dd98a: Status 404 returned error can't find the container with id c2ad74d32f8657dd09a746af4d845661390188b07f6a65c51c1ef06af45dd98a Sep 29 21:13:14 crc kubenswrapper[4741]: I0929 21:13:14.071866 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" event={"ID":"f372c5ae-6094-4c36-8b23-0a06a4e50c93","Type":"ContainerStarted","Data":"5405ddc81e5caed2372a712ec363dda59f35fd14075edae448439e61f5170cb0"} Sep 29 21:13:14 crc kubenswrapper[4741]: I0929 21:13:14.072145 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" event={"ID":"f372c5ae-6094-4c36-8b23-0a06a4e50c93","Type":"ContainerStarted","Data":"c2ad74d32f8657dd09a746af4d845661390188b07f6a65c51c1ef06af45dd98a"} Sep 29 21:13:14 crc kubenswrapper[4741]: I0929 21:13:14.226159 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:14 crc kubenswrapper[4741]: I0929 21:13:14.228936 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:14 crc kubenswrapper[4741]: I0929 21:13:14.280670 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:14 crc kubenswrapper[4741]: I0929 21:13:14.296909 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" podStartSLOduration=1.805790153 podStartE2EDuration="2.296890508s" podCreationTimestamp="2025-09-29 21:13:12 +0000 UTC" firstStartedPulling="2025-09-29 21:13:13.090679858 +0000 UTC m=+7434.738469200" lastFinishedPulling="2025-09-29 21:13:13.581780183 +0000 UTC m=+7435.229569555" observedRunningTime="2025-09-29 21:13:14.095850346 +0000 UTC m=+7435.743639718" watchObservedRunningTime="2025-09-29 21:13:14.296890508 +0000 UTC m=+7435.944679840" Sep 29 21:13:15 crc kubenswrapper[4741]: I0929 21:13:15.144975 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:15 crc kubenswrapper[4741]: I0929 21:13:15.239133 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbqfj"] Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.101264 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wbqfj" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="registry-server" containerID="cri-o://b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7" gracePeriod=2 Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.684201 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.809597 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-utilities\") pod \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.809881 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbjmr\" (UniqueName: \"kubernetes.io/projected/aa4eb3c5-e143-4aec-8cd6-6446638f4900-kube-api-access-mbjmr\") pod \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.810268 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-catalog-content\") pod \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\" (UID: \"aa4eb3c5-e143-4aec-8cd6-6446638f4900\") " Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.811686 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-utilities" (OuterVolumeSpecName: "utilities") pod "aa4eb3c5-e143-4aec-8cd6-6446638f4900" (UID: "aa4eb3c5-e143-4aec-8cd6-6446638f4900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.812742 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.816743 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa4eb3c5-e143-4aec-8cd6-6446638f4900-kube-api-access-mbjmr" (OuterVolumeSpecName: "kube-api-access-mbjmr") pod "aa4eb3c5-e143-4aec-8cd6-6446638f4900" (UID: "aa4eb3c5-e143-4aec-8cd6-6446638f4900"). InnerVolumeSpecName "kube-api-access-mbjmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.890597 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa4eb3c5-e143-4aec-8cd6-6446638f4900" (UID: "aa4eb3c5-e143-4aec-8cd6-6446638f4900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.915279 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbjmr\" (UniqueName: \"kubernetes.io/projected/aa4eb3c5-e143-4aec-8cd6-6446638f4900-kube-api-access-mbjmr\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:17 crc kubenswrapper[4741]: I0929 21:13:17.915337 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4eb3c5-e143-4aec-8cd6-6446638f4900-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.114831 4741 generic.go:334] "Generic (PLEG): container finished" podID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerID="b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7" exitCode=0 Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.114909 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqfj" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.114882 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqfj" event={"ID":"aa4eb3c5-e143-4aec-8cd6-6446638f4900","Type":"ContainerDied","Data":"b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7"} Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.114991 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqfj" event={"ID":"aa4eb3c5-e143-4aec-8cd6-6446638f4900","Type":"ContainerDied","Data":"67db57a858a054b555e9466b0233c45bc7d8cbf62dd3a00ac6f0862cf61a239b"} Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.115024 4741 scope.go:117] "RemoveContainer" containerID="b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.156667 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbqfj"] Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.157495 4741 scope.go:117] "RemoveContainer" containerID="669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.179860 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wbqfj"] Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.196306 4741 scope.go:117] "RemoveContainer" containerID="b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.253820 4741 scope.go:117] "RemoveContainer" containerID="b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7" Sep 29 21:13:18 crc kubenswrapper[4741]: E0929 21:13:18.254315 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7\": container with ID starting with b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7 not found: ID does not exist" containerID="b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.254346 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7"} err="failed to get container status \"b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7\": rpc error: code = NotFound desc = could not find container \"b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7\": container with ID starting with b0776485b4e7cd213e3460fbb3373afe9304daabf692710ea41f3cb8b7eee2b7 not found: ID does not exist" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.254380 4741 scope.go:117] "RemoveContainer" containerID="669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46" Sep 29 21:13:18 crc kubenswrapper[4741]: E0929 21:13:18.254664 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46\": container with ID starting with 669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46 not found: ID does not exist" containerID="669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.254691 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46"} err="failed to get container status \"669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46\": rpc error: code = NotFound desc = could not find container \"669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46\": container with ID starting with 669f4b66d6f27da26d1e18b4143e59922824825e21896f33ee19e74e38a07f46 not found: ID does not exist" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.254708 4741 scope.go:117] "RemoveContainer" containerID="b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1" Sep 29 21:13:18 crc kubenswrapper[4741]: E0929 21:13:18.254920 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1\": container with ID starting with b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1 not found: ID does not exist" containerID="b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1" Sep 29 21:13:18 crc kubenswrapper[4741]: I0929 21:13:18.254945 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1"} err="failed to get container status \"b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1\": rpc error: code = NotFound desc = could not find container \"b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1\": container with ID starting with b65318f5f13481960c490e3315599b275081523ffdd028848a8180724545e9c1 not found: ID does not exist" Sep 29 21:13:19 crc kubenswrapper[4741]: I0929 21:13:19.112555 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" path="/var/lib/kubelet/pods/aa4eb3c5-e143-4aec-8cd6-6446638f4900/volumes" Sep 29 21:13:26 crc kubenswrapper[4741]: I0929 21:13:26.085616 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:13:26 crc kubenswrapper[4741]: E0929 21:13:26.086571 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:13:40 crc kubenswrapper[4741]: I0929 21:13:40.086322 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:13:40 crc kubenswrapper[4741]: E0929 21:13:40.087198 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:13:53 crc kubenswrapper[4741]: I0929 21:13:53.085433 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:13:53 crc kubenswrapper[4741]: E0929 21:13:53.086138 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:13:57 crc kubenswrapper[4741]: I0929 21:13:57.545023 4741 generic.go:334] "Generic (PLEG): container finished" podID="f372c5ae-6094-4c36-8b23-0a06a4e50c93" containerID="5405ddc81e5caed2372a712ec363dda59f35fd14075edae448439e61f5170cb0" exitCode=0 Sep 29 21:13:57 crc kubenswrapper[4741]: I0929 21:13:57.545616 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" event={"ID":"f372c5ae-6094-4c36-8b23-0a06a4e50c93","Type":"ContainerDied","Data":"5405ddc81e5caed2372a712ec363dda59f35fd14075edae448439e61f5170cb0"} Sep 29 21:13:58 crc kubenswrapper[4741]: I0929 21:13:58.983539 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.117863 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-inventory\") pod \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.117916 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ceph\") pod \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.118126 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw9nv\" (UniqueName: \"kubernetes.io/projected/f372c5ae-6094-4c36-8b23-0a06a4e50c93-kube-api-access-xw9nv\") pod \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.118480 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ssh-key\") pod \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\" (UID: \"f372c5ae-6094-4c36-8b23-0a06a4e50c93\") " Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.123815 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ceph" (OuterVolumeSpecName: "ceph") pod "f372c5ae-6094-4c36-8b23-0a06a4e50c93" (UID: "f372c5ae-6094-4c36-8b23-0a06a4e50c93"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.124481 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f372c5ae-6094-4c36-8b23-0a06a4e50c93-kube-api-access-xw9nv" (OuterVolumeSpecName: "kube-api-access-xw9nv") pod "f372c5ae-6094-4c36-8b23-0a06a4e50c93" (UID: "f372c5ae-6094-4c36-8b23-0a06a4e50c93"). InnerVolumeSpecName "kube-api-access-xw9nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.154173 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-inventory" (OuterVolumeSpecName: "inventory") pod "f372c5ae-6094-4c36-8b23-0a06a4e50c93" (UID: "f372c5ae-6094-4c36-8b23-0a06a4e50c93"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.154212 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f372c5ae-6094-4c36-8b23-0a06a4e50c93" (UID: "f372c5ae-6094-4c36-8b23-0a06a4e50c93"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.221129 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw9nv\" (UniqueName: \"kubernetes.io/projected/f372c5ae-6094-4c36-8b23-0a06a4e50c93-kube-api-access-xw9nv\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.221162 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.221187 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.221199 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f372c5ae-6094-4c36-8b23-0a06a4e50c93-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.564841 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" event={"ID":"f372c5ae-6094-4c36-8b23-0a06a4e50c93","Type":"ContainerDied","Data":"c2ad74d32f8657dd09a746af4d845661390188b07f6a65c51c1ef06af45dd98a"} Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.565126 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2ad74d32f8657dd09a746af4d845661390188b07f6a65c51c1ef06af45dd98a" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.564895 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hwlcx" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.641743 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-c9bsh"] Sep 29 21:13:59 crc kubenswrapper[4741]: E0929 21:13:59.642170 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f372c5ae-6094-4c36-8b23-0a06a4e50c93" containerName="configure-os-openstack-openstack-cell1" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.642192 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f372c5ae-6094-4c36-8b23-0a06a4e50c93" containerName="configure-os-openstack-openstack-cell1" Sep 29 21:13:59 crc kubenswrapper[4741]: E0929 21:13:59.642231 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="extract-utilities" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.642240 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="extract-utilities" Sep 29 21:13:59 crc kubenswrapper[4741]: E0929 21:13:59.642249 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="registry-server" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.642257 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="registry-server" Sep 29 21:13:59 crc kubenswrapper[4741]: E0929 21:13:59.642268 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="extract-content" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.642276 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="extract-content" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.642546 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa4eb3c5-e143-4aec-8cd6-6446638f4900" containerName="registry-server" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.642596 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f372c5ae-6094-4c36-8b23-0a06a4e50c93" containerName="configure-os-openstack-openstack-cell1" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.643500 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.645648 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.645697 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.645704 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.645872 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.653988 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-c9bsh"] Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.732697 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-inventory-0\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.732761 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfm7w\" (UniqueName: \"kubernetes.io/projected/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-kube-api-access-mfm7w\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.732948 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.733189 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ceph\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.834647 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ceph\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.834749 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-inventory-0\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.834780 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfm7w\" (UniqueName: \"kubernetes.io/projected/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-kube-api-access-mfm7w\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.834822 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.838555 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.839052 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ceph\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.841811 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-inventory-0\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.852606 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfm7w\" (UniqueName: \"kubernetes.io/projected/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-kube-api-access-mfm7w\") pod \"ssh-known-hosts-openstack-c9bsh\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:13:59 crc kubenswrapper[4741]: I0929 21:13:59.969905 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:14:00 crc kubenswrapper[4741]: I0929 21:14:00.519650 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-c9bsh"] Sep 29 21:14:00 crc kubenswrapper[4741]: W0929 21:14:00.523794 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9b60bc5_0ab8_46c2_adeb_158cc32b9982.slice/crio-44a8820ddef83682102a9045e79d63a5d0a746659b3e8668e8fd10f14959f8aa WatchSource:0}: Error finding container 44a8820ddef83682102a9045e79d63a5d0a746659b3e8668e8fd10f14959f8aa: Status 404 returned error can't find the container with id 44a8820ddef83682102a9045e79d63a5d0a746659b3e8668e8fd10f14959f8aa Sep 29 21:14:00 crc kubenswrapper[4741]: I0929 21:14:00.577241 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-c9bsh" event={"ID":"b9b60bc5-0ab8-46c2-adeb-158cc32b9982","Type":"ContainerStarted","Data":"44a8820ddef83682102a9045e79d63a5d0a746659b3e8668e8fd10f14959f8aa"} Sep 29 21:14:01 crc kubenswrapper[4741]: I0929 21:14:01.588116 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-c9bsh" event={"ID":"b9b60bc5-0ab8-46c2-adeb-158cc32b9982","Type":"ContainerStarted","Data":"dac2b73e5dda3d50c096eedc0a20508f7d0850e5167da9ffff0253299c3298a7"} Sep 29 21:14:01 crc kubenswrapper[4741]: I0929 21:14:01.613100 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-c9bsh" podStartSLOduration=2.033140768 podStartE2EDuration="2.613033595s" podCreationTimestamp="2025-09-29 21:13:59 +0000 UTC" firstStartedPulling="2025-09-29 21:14:00.526969811 +0000 UTC m=+7482.174759143" lastFinishedPulling="2025-09-29 21:14:01.106862638 +0000 UTC m=+7482.754651970" observedRunningTime="2025-09-29 21:14:01.608922895 +0000 UTC m=+7483.256712237" watchObservedRunningTime="2025-09-29 21:14:01.613033595 +0000 UTC m=+7483.260822927" Sep 29 21:14:08 crc kubenswrapper[4741]: I0929 21:14:08.086117 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:14:08 crc kubenswrapper[4741]: E0929 21:14:08.086962 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:14:10 crc kubenswrapper[4741]: I0929 21:14:10.676779 4741 generic.go:334] "Generic (PLEG): container finished" podID="b9b60bc5-0ab8-46c2-adeb-158cc32b9982" containerID="dac2b73e5dda3d50c096eedc0a20508f7d0850e5167da9ffff0253299c3298a7" exitCode=0 Sep 29 21:14:10 crc kubenswrapper[4741]: I0929 21:14:10.676898 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-c9bsh" event={"ID":"b9b60bc5-0ab8-46c2-adeb-158cc32b9982","Type":"ContainerDied","Data":"dac2b73e5dda3d50c096eedc0a20508f7d0850e5167da9ffff0253299c3298a7"} Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.126547 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.198815 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-inventory-0\") pod \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.198862 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ssh-key-openstack-cell1\") pod \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.198897 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfm7w\" (UniqueName: \"kubernetes.io/projected/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-kube-api-access-mfm7w\") pod \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.199019 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ceph\") pod \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\" (UID: \"b9b60bc5-0ab8-46c2-adeb-158cc32b9982\") " Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.206275 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-kube-api-access-mfm7w" (OuterVolumeSpecName: "kube-api-access-mfm7w") pod "b9b60bc5-0ab8-46c2-adeb-158cc32b9982" (UID: "b9b60bc5-0ab8-46c2-adeb-158cc32b9982"). InnerVolumeSpecName "kube-api-access-mfm7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.206594 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ceph" (OuterVolumeSpecName: "ceph") pod "b9b60bc5-0ab8-46c2-adeb-158cc32b9982" (UID: "b9b60bc5-0ab8-46c2-adeb-158cc32b9982"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.230890 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b9b60bc5-0ab8-46c2-adeb-158cc32b9982" (UID: "b9b60bc5-0ab8-46c2-adeb-158cc32b9982"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.238930 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b9b60bc5-0ab8-46c2-adeb-158cc32b9982" (UID: "b9b60bc5-0ab8-46c2-adeb-158cc32b9982"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.301565 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.301604 4741 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-inventory-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.301621 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.301636 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfm7w\" (UniqueName: \"kubernetes.io/projected/b9b60bc5-0ab8-46c2-adeb-158cc32b9982-kube-api-access-mfm7w\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.695860 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-c9bsh" event={"ID":"b9b60bc5-0ab8-46c2-adeb-158cc32b9982","Type":"ContainerDied","Data":"44a8820ddef83682102a9045e79d63a5d0a746659b3e8668e8fd10f14959f8aa"} Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.695908 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44a8820ddef83682102a9045e79d63a5d0a746659b3e8668e8fd10f14959f8aa" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.695920 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-c9bsh" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.776341 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-7jxgv"] Sep 29 21:14:12 crc kubenswrapper[4741]: E0929 21:14:12.777821 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b60bc5-0ab8-46c2-adeb-158cc32b9982" containerName="ssh-known-hosts-openstack" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.777855 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b60bc5-0ab8-46c2-adeb-158cc32b9982" containerName="ssh-known-hosts-openstack" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.778243 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b60bc5-0ab8-46c2-adeb-158cc32b9982" containerName="ssh-known-hosts-openstack" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.780120 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.785260 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.785491 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.785673 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.785776 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.785904 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-7jxgv"] Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.812255 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ssh-key\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.812502 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ceph\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.812681 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hwdk\" (UniqueName: \"kubernetes.io/projected/7dfada48-0466-4d17-9ef0-a6ffc08f1035-kube-api-access-6hwdk\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.812811 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-inventory\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.914212 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-inventory\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.914299 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ssh-key\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.914375 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ceph\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.914531 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hwdk\" (UniqueName: \"kubernetes.io/projected/7dfada48-0466-4d17-9ef0-a6ffc08f1035-kube-api-access-6hwdk\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.918966 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-inventory\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.919238 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ceph\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.923215 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ssh-key\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:12 crc kubenswrapper[4741]: I0929 21:14:12.931324 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hwdk\" (UniqueName: \"kubernetes.io/projected/7dfada48-0466-4d17-9ef0-a6ffc08f1035-kube-api-access-6hwdk\") pod \"run-os-openstack-openstack-cell1-7jxgv\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:13 crc kubenswrapper[4741]: I0929 21:14:13.112087 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:13 crc kubenswrapper[4741]: I0929 21:14:13.679169 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-7jxgv"] Sep 29 21:14:13 crc kubenswrapper[4741]: I0929 21:14:13.710939 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" event={"ID":"7dfada48-0466-4d17-9ef0-a6ffc08f1035","Type":"ContainerStarted","Data":"a7b9e577d0ac25128161d1f75333a66b50ca775815c2c7018555bc1d3878b1bc"} Sep 29 21:14:15 crc kubenswrapper[4741]: I0929 21:14:15.731112 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" event={"ID":"7dfada48-0466-4d17-9ef0-a6ffc08f1035","Type":"ContainerStarted","Data":"950a67db4c4ca65cf3144482fe93d1ca7954c5c55f18e960f50a28ad573e3a31"} Sep 29 21:14:15 crc kubenswrapper[4741]: I0929 21:14:15.752307 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" podStartSLOduration=2.488848796 podStartE2EDuration="3.7522891s" podCreationTimestamp="2025-09-29 21:14:12 +0000 UTC" firstStartedPulling="2025-09-29 21:14:13.691847313 +0000 UTC m=+7495.339636645" lastFinishedPulling="2025-09-29 21:14:14.955287617 +0000 UTC m=+7496.603076949" observedRunningTime="2025-09-29 21:14:15.744429693 +0000 UTC m=+7497.392219025" watchObservedRunningTime="2025-09-29 21:14:15.7522891 +0000 UTC m=+7497.400078432" Sep 29 21:14:20 crc kubenswrapper[4741]: I0929 21:14:20.086742 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:14:20 crc kubenswrapper[4741]: E0929 21:14:20.087657 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:14:22 crc kubenswrapper[4741]: I0929 21:14:22.797269 4741 generic.go:334] "Generic (PLEG): container finished" podID="7dfada48-0466-4d17-9ef0-a6ffc08f1035" containerID="950a67db4c4ca65cf3144482fe93d1ca7954c5c55f18e960f50a28ad573e3a31" exitCode=0 Sep 29 21:14:22 crc kubenswrapper[4741]: I0929 21:14:22.797446 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" event={"ID":"7dfada48-0466-4d17-9ef0-a6ffc08f1035","Type":"ContainerDied","Data":"950a67db4c4ca65cf3144482fe93d1ca7954c5c55f18e960f50a28ad573e3a31"} Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.234810 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.244268 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-inventory\") pod \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.244415 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ceph\") pod \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.244438 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hwdk\" (UniqueName: \"kubernetes.io/projected/7dfada48-0466-4d17-9ef0-a6ffc08f1035-kube-api-access-6hwdk\") pod \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.253252 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfada48-0466-4d17-9ef0-a6ffc08f1035-kube-api-access-6hwdk" (OuterVolumeSpecName: "kube-api-access-6hwdk") pod "7dfada48-0466-4d17-9ef0-a6ffc08f1035" (UID: "7dfada48-0466-4d17-9ef0-a6ffc08f1035"). InnerVolumeSpecName "kube-api-access-6hwdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.255258 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ceph" (OuterVolumeSpecName: "ceph") pod "7dfada48-0466-4d17-9ef0-a6ffc08f1035" (UID: "7dfada48-0466-4d17-9ef0-a6ffc08f1035"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.282223 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-inventory" (OuterVolumeSpecName: "inventory") pod "7dfada48-0466-4d17-9ef0-a6ffc08f1035" (UID: "7dfada48-0466-4d17-9ef0-a6ffc08f1035"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.345887 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ssh-key\") pod \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\" (UID: \"7dfada48-0466-4d17-9ef0-a6ffc08f1035\") " Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.346682 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.346705 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hwdk\" (UniqueName: \"kubernetes.io/projected/7dfada48-0466-4d17-9ef0-a6ffc08f1035-kube-api-access-6hwdk\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.346717 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.376492 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7dfada48-0466-4d17-9ef0-a6ffc08f1035" (UID: "7dfada48-0466-4d17-9ef0-a6ffc08f1035"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.448523 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dfada48-0466-4d17-9ef0-a6ffc08f1035-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.814926 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" event={"ID":"7dfada48-0466-4d17-9ef0-a6ffc08f1035","Type":"ContainerDied","Data":"a7b9e577d0ac25128161d1f75333a66b50ca775815c2c7018555bc1d3878b1bc"} Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.814963 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7jxgv" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.814964 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b9e577d0ac25128161d1f75333a66b50ca775815c2c7018555bc1d3878b1bc" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.889727 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-562bz"] Sep 29 21:14:24 crc kubenswrapper[4741]: E0929 21:14:24.890617 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfada48-0466-4d17-9ef0-a6ffc08f1035" containerName="run-os-openstack-openstack-cell1" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.890643 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfada48-0466-4d17-9ef0-a6ffc08f1035" containerName="run-os-openstack-openstack-cell1" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.890920 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dfada48-0466-4d17-9ef0-a6ffc08f1035" containerName="run-os-openstack-openstack-cell1" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.892022 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.894184 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.894517 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.894675 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.895137 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:14:24 crc kubenswrapper[4741]: I0929 21:14:24.920868 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-562bz"] Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.060950 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ceph\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.061075 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc22w\" (UniqueName: \"kubernetes.io/projected/fb553e7a-d9f7-4eb9-8aba-8452d128727c-kube-api-access-lc22w\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.061106 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.061175 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-inventory\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.162875 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ceph\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.162965 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc22w\" (UniqueName: \"kubernetes.io/projected/fb553e7a-d9f7-4eb9-8aba-8452d128727c-kube-api-access-lc22w\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.162987 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.163035 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-inventory\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.167100 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-inventory\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.167465 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ceph\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.178896 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.179715 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc22w\" (UniqueName: \"kubernetes.io/projected/fb553e7a-d9f7-4eb9-8aba-8452d128727c-kube-api-access-lc22w\") pod \"reboot-os-openstack-openstack-cell1-562bz\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.208105 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.738765 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-562bz"] Sep 29 21:14:25 crc kubenswrapper[4741]: I0929 21:14:25.824241 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" event={"ID":"fb553e7a-d9f7-4eb9-8aba-8452d128727c","Type":"ContainerStarted","Data":"52302de220cd06ddc9f602ac6776d81c7cc8f15903b517161b47a97180b1eb1b"} Sep 29 21:14:27 crc kubenswrapper[4741]: I0929 21:14:27.848308 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" event={"ID":"fb553e7a-d9f7-4eb9-8aba-8452d128727c","Type":"ContainerStarted","Data":"59ab272ea5012b12e8430aaab578d2230156f2689df4e56bb3373f94da3c24f1"} Sep 29 21:14:27 crc kubenswrapper[4741]: I0929 21:14:27.870598 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" podStartSLOduration=2.435133251 podStartE2EDuration="3.870579956s" podCreationTimestamp="2025-09-29 21:14:24 +0000 UTC" firstStartedPulling="2025-09-29 21:14:25.745908116 +0000 UTC m=+7507.393697448" lastFinishedPulling="2025-09-29 21:14:27.181354821 +0000 UTC m=+7508.829144153" observedRunningTime="2025-09-29 21:14:27.866162457 +0000 UTC m=+7509.513951879" watchObservedRunningTime="2025-09-29 21:14:27.870579956 +0000 UTC m=+7509.518369288" Sep 29 21:14:35 crc kubenswrapper[4741]: I0929 21:14:35.086842 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:14:35 crc kubenswrapper[4741]: E0929 21:14:35.087744 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:14:43 crc kubenswrapper[4741]: I0929 21:14:43.017104 4741 generic.go:334] "Generic (PLEG): container finished" podID="fb553e7a-d9f7-4eb9-8aba-8452d128727c" containerID="59ab272ea5012b12e8430aaab578d2230156f2689df4e56bb3373f94da3c24f1" exitCode=0 Sep 29 21:14:43 crc kubenswrapper[4741]: I0929 21:14:43.017194 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" event={"ID":"fb553e7a-d9f7-4eb9-8aba-8452d128727c","Type":"ContainerDied","Data":"59ab272ea5012b12e8430aaab578d2230156f2689df4e56bb3373f94da3c24f1"} Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.548562 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.690289 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-inventory\") pod \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.690347 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc22w\" (UniqueName: \"kubernetes.io/projected/fb553e7a-d9f7-4eb9-8aba-8452d128727c-kube-api-access-lc22w\") pod \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.690519 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ssh-key\") pod \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.690550 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ceph\") pod \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\" (UID: \"fb553e7a-d9f7-4eb9-8aba-8452d128727c\") " Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.697562 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ceph" (OuterVolumeSpecName: "ceph") pod "fb553e7a-d9f7-4eb9-8aba-8452d128727c" (UID: "fb553e7a-d9f7-4eb9-8aba-8452d128727c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.698891 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb553e7a-d9f7-4eb9-8aba-8452d128727c-kube-api-access-lc22w" (OuterVolumeSpecName: "kube-api-access-lc22w") pod "fb553e7a-d9f7-4eb9-8aba-8452d128727c" (UID: "fb553e7a-d9f7-4eb9-8aba-8452d128727c"). InnerVolumeSpecName "kube-api-access-lc22w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.720043 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb553e7a-d9f7-4eb9-8aba-8452d128727c" (UID: "fb553e7a-d9f7-4eb9-8aba-8452d128727c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.726889 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-inventory" (OuterVolumeSpecName: "inventory") pod "fb553e7a-d9f7-4eb9-8aba-8452d128727c" (UID: "fb553e7a-d9f7-4eb9-8aba-8452d128727c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.797669 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.797716 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.797735 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb553e7a-d9f7-4eb9-8aba-8452d128727c-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:44 crc kubenswrapper[4741]: I0929 21:14:44.797753 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc22w\" (UniqueName: \"kubernetes.io/projected/fb553e7a-d9f7-4eb9-8aba-8452d128727c-kube-api-access-lc22w\") on node \"crc\" DevicePath \"\"" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.032778 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" event={"ID":"fb553e7a-d9f7-4eb9-8aba-8452d128727c","Type":"ContainerDied","Data":"52302de220cd06ddc9f602ac6776d81c7cc8f15903b517161b47a97180b1eb1b"} Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.032816 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52302de220cd06ddc9f602ac6776d81c7cc8f15903b517161b47a97180b1eb1b" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.033225 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-562bz" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.155275 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-tl5j6"] Sep 29 21:14:45 crc kubenswrapper[4741]: E0929 21:14:45.202694 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb553e7a-d9f7-4eb9-8aba-8452d128727c" containerName="reboot-os-openstack-openstack-cell1" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.202733 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb553e7a-d9f7-4eb9-8aba-8452d128727c" containerName="reboot-os-openstack-openstack-cell1" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.203656 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb553e7a-d9f7-4eb9-8aba-8452d128727c" containerName="reboot-os-openstack-openstack-cell1" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.204676 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-tl5j6"] Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.204853 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.208079 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.208247 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.208273 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-inventory\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.208377 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.208443 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctbnt\" (UniqueName: \"kubernetes.io/projected/5d41613b-ad82-41fa-893a-e150f9eccb58-kube-api-access-ctbnt\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.209142 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.209519 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.209624 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.209645 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.210080 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ceph\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.210138 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.210171 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.210248 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.210301 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ssh-key\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.210359 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.211789 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.313871 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.313918 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ssh-key\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.313954 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.313981 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314024 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314064 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314080 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-inventory\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314115 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314139 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctbnt\" (UniqueName: \"kubernetes.io/projected/5d41613b-ad82-41fa-893a-e150f9eccb58-kube-api-access-ctbnt\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314184 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ceph\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314205 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.314223 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.317955 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.318321 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.318896 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ssh-key\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.319172 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ceph\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.319428 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-inventory\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.319527 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.319847 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.319939 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.320436 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.322190 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.327023 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.330451 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctbnt\" (UniqueName: \"kubernetes.io/projected/5d41613b-ad82-41fa-893a-e150f9eccb58-kube-api-access-ctbnt\") pod \"install-certs-openstack-openstack-cell1-tl5j6\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:45 crc kubenswrapper[4741]: I0929 21:14:45.526681 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:14:46 crc kubenswrapper[4741]: I0929 21:14:46.059777 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-tl5j6"] Sep 29 21:14:46 crc kubenswrapper[4741]: I0929 21:14:46.086775 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:14:46 crc kubenswrapper[4741]: E0929 21:14:46.087125 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:14:47 crc kubenswrapper[4741]: I0929 21:14:47.055873 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" event={"ID":"5d41613b-ad82-41fa-893a-e150f9eccb58","Type":"ContainerStarted","Data":"44ad8ef6146c067d0f0e67827b59a07ca9edcfd8572a394776315bb6adf7f3df"} Sep 29 21:14:48 crc kubenswrapper[4741]: I0929 21:14:48.067577 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" event={"ID":"5d41613b-ad82-41fa-893a-e150f9eccb58","Type":"ContainerStarted","Data":"cb99aebeebf2909284f4bd02d99daf267ee128bf0b33b65fb116e74866756e49"} Sep 29 21:14:48 crc kubenswrapper[4741]: I0929 21:14:48.095139 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" podStartSLOduration=1.772583171 podStartE2EDuration="3.095117767s" podCreationTimestamp="2025-09-29 21:14:45 +0000 UTC" firstStartedPulling="2025-09-29 21:14:46.064040821 +0000 UTC m=+7527.711830153" lastFinishedPulling="2025-09-29 21:14:47.386575417 +0000 UTC m=+7529.034364749" observedRunningTime="2025-09-29 21:14:48.092892807 +0000 UTC m=+7529.740682149" watchObservedRunningTime="2025-09-29 21:14:48.095117767 +0000 UTC m=+7529.742907129" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.136897 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl"] Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.141724 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.148778 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.149384 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.177695 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl"] Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.240496 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-config-volume\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.240716 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-secret-volume\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.241182 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbmrv\" (UniqueName: \"kubernetes.io/projected/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-kube-api-access-lbmrv\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.343723 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-secret-volume\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.343825 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbmrv\" (UniqueName: \"kubernetes.io/projected/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-kube-api-access-lbmrv\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.343859 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-config-volume\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.344878 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-config-volume\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.352191 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-secret-volume\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.360836 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbmrv\" (UniqueName: \"kubernetes.io/projected/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-kube-api-access-lbmrv\") pod \"collect-profiles-29319675-xjtpl\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:00 crc kubenswrapper[4741]: I0929 21:15:00.474821 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:02 crc kubenswrapper[4741]: W0929 21:15:00.929712 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98ab01c7_5aab_45fd_b45d_1e8508dd0c16.slice/crio-ecf3f4b265f577e59360e595f9362fd3348d99fa89bfc5da820c0972bc1114fc WatchSource:0}: Error finding container ecf3f4b265f577e59360e595f9362fd3348d99fa89bfc5da820c0972bc1114fc: Status 404 returned error can't find the container with id ecf3f4b265f577e59360e595f9362fd3348d99fa89bfc5da820c0972bc1114fc Sep 29 21:15:02 crc kubenswrapper[4741]: I0929 21:15:00.935866 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl"] Sep 29 21:15:02 crc kubenswrapper[4741]: I0929 21:15:01.086051 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:15:02 crc kubenswrapper[4741]: E0929 21:15:01.086295 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:15:02 crc kubenswrapper[4741]: I0929 21:15:01.227231 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" event={"ID":"98ab01c7-5aab-45fd-b45d-1e8508dd0c16","Type":"ContainerStarted","Data":"694ffb09050eca501cdc4eed77bfc64b4b536ba062c1a04e40b53486cdbc74da"} Sep 29 21:15:02 crc kubenswrapper[4741]: I0929 21:15:01.227271 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" event={"ID":"98ab01c7-5aab-45fd-b45d-1e8508dd0c16","Type":"ContainerStarted","Data":"ecf3f4b265f577e59360e595f9362fd3348d99fa89bfc5da820c0972bc1114fc"} Sep 29 21:15:02 crc kubenswrapper[4741]: I0929 21:15:01.250210 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" podStartSLOduration=1.250167701 podStartE2EDuration="1.250167701s" podCreationTimestamp="2025-09-29 21:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:15:01.242291064 +0000 UTC m=+7542.890080396" watchObservedRunningTime="2025-09-29 21:15:01.250167701 +0000 UTC m=+7542.897957033" Sep 29 21:15:03 crc kubenswrapper[4741]: I0929 21:15:03.248076 4741 generic.go:334] "Generic (PLEG): container finished" podID="98ab01c7-5aab-45fd-b45d-1e8508dd0c16" containerID="694ffb09050eca501cdc4eed77bfc64b4b536ba062c1a04e40b53486cdbc74da" exitCode=0 Sep 29 21:15:03 crc kubenswrapper[4741]: I0929 21:15:03.248171 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" event={"ID":"98ab01c7-5aab-45fd-b45d-1e8508dd0c16","Type":"ContainerDied","Data":"694ffb09050eca501cdc4eed77bfc64b4b536ba062c1a04e40b53486cdbc74da"} Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.658598 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.732065 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-config-volume\") pod \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.732131 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbmrv\" (UniqueName: \"kubernetes.io/projected/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-kube-api-access-lbmrv\") pod \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.732230 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-secret-volume\") pod \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\" (UID: \"98ab01c7-5aab-45fd-b45d-1e8508dd0c16\") " Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.732904 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-config-volume" (OuterVolumeSpecName: "config-volume") pod "98ab01c7-5aab-45fd-b45d-1e8508dd0c16" (UID: "98ab01c7-5aab-45fd-b45d-1e8508dd0c16"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.737645 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "98ab01c7-5aab-45fd-b45d-1e8508dd0c16" (UID: "98ab01c7-5aab-45fd-b45d-1e8508dd0c16"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.738597 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-kube-api-access-lbmrv" (OuterVolumeSpecName: "kube-api-access-lbmrv") pod "98ab01c7-5aab-45fd-b45d-1e8508dd0c16" (UID: "98ab01c7-5aab-45fd-b45d-1e8508dd0c16"). InnerVolumeSpecName "kube-api-access-lbmrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.834691 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.834725 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbmrv\" (UniqueName: \"kubernetes.io/projected/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-kube-api-access-lbmrv\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:04 crc kubenswrapper[4741]: I0929 21:15:04.834737 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98ab01c7-5aab-45fd-b45d-1e8508dd0c16-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:05 crc kubenswrapper[4741]: I0929 21:15:05.270058 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" event={"ID":"98ab01c7-5aab-45fd-b45d-1e8508dd0c16","Type":"ContainerDied","Data":"ecf3f4b265f577e59360e595f9362fd3348d99fa89bfc5da820c0972bc1114fc"} Sep 29 21:15:05 crc kubenswrapper[4741]: I0929 21:15:05.270401 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecf3f4b265f577e59360e595f9362fd3348d99fa89bfc5da820c0972bc1114fc" Sep 29 21:15:05 crc kubenswrapper[4741]: I0929 21:15:05.270086 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl" Sep 29 21:15:05 crc kubenswrapper[4741]: I0929 21:15:05.319645 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn"] Sep 29 21:15:05 crc kubenswrapper[4741]: I0929 21:15:05.327826 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319630-cgswn"] Sep 29 21:15:06 crc kubenswrapper[4741]: I0929 21:15:06.279656 4741 generic.go:334] "Generic (PLEG): container finished" podID="5d41613b-ad82-41fa-893a-e150f9eccb58" containerID="cb99aebeebf2909284f4bd02d99daf267ee128bf0b33b65fb116e74866756e49" exitCode=0 Sep 29 21:15:06 crc kubenswrapper[4741]: I0929 21:15:06.279699 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" event={"ID":"5d41613b-ad82-41fa-893a-e150f9eccb58","Type":"ContainerDied","Data":"cb99aebeebf2909284f4bd02d99daf267ee128bf0b33b65fb116e74866756e49"} Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.099285 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75859263-068b-4079-93f2-f4015b2c5034" path="/var/lib/kubelet/pods/75859263-068b-4079-93f2-f4015b2c5034/volumes" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.733746 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808084 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-inventory\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808136 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-nova-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808204 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctbnt\" (UniqueName: \"kubernetes.io/projected/5d41613b-ad82-41fa-893a-e150f9eccb58-kube-api-access-ctbnt\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808270 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-dhcp-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808301 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-bootstrap-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808320 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ovn-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808335 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ceph\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808361 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-metadata-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808435 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-telemetry-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808474 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-libvirt-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808530 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ssh-key\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.808579 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-sriov-combined-ca-bundle\") pod \"5d41613b-ad82-41fa-893a-e150f9eccb58\" (UID: \"5d41613b-ad82-41fa-893a-e150f9eccb58\") " Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.813361 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.813685 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.815710 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.817812 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ceph" (OuterVolumeSpecName: "ceph") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.818300 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.819620 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d41613b-ad82-41fa-893a-e150f9eccb58-kube-api-access-ctbnt" (OuterVolumeSpecName: "kube-api-access-ctbnt") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "kube-api-access-ctbnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.820921 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.822808 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.824925 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.825657 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.857819 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.870704 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-inventory" (OuterVolumeSpecName: "inventory") pod "5d41613b-ad82-41fa-893a-e150f9eccb58" (UID: "5d41613b-ad82-41fa-893a-e150f9eccb58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911098 4741 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911133 4741 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911143 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911151 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911161 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911170 4741 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911179 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctbnt\" (UniqueName: \"kubernetes.io/projected/5d41613b-ad82-41fa-893a-e150f9eccb58-kube-api-access-ctbnt\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911191 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911202 4741 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911214 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911224 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:07 crc kubenswrapper[4741]: I0929 21:15:07.911234 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d41613b-ad82-41fa-893a-e150f9eccb58-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.299167 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" event={"ID":"5d41613b-ad82-41fa-893a-e150f9eccb58","Type":"ContainerDied","Data":"44ad8ef6146c067d0f0e67827b59a07ca9edcfd8572a394776315bb6adf7f3df"} Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.299495 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44ad8ef6146c067d0f0e67827b59a07ca9edcfd8572a394776315bb6adf7f3df" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.299231 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tl5j6" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.401561 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-n5fvd"] Sep 29 21:15:08 crc kubenswrapper[4741]: E0929 21:15:08.402032 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ab01c7-5aab-45fd-b45d-1e8508dd0c16" containerName="collect-profiles" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.402043 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ab01c7-5aab-45fd-b45d-1e8508dd0c16" containerName="collect-profiles" Sep 29 21:15:08 crc kubenswrapper[4741]: E0929 21:15:08.402097 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d41613b-ad82-41fa-893a-e150f9eccb58" containerName="install-certs-openstack-openstack-cell1" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.402103 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d41613b-ad82-41fa-893a-e150f9eccb58" containerName="install-certs-openstack-openstack-cell1" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.402319 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d41613b-ad82-41fa-893a-e150f9eccb58" containerName="install-certs-openstack-openstack-cell1" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.402342 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ab01c7-5aab-45fd-b45d-1e8508dd0c16" containerName="collect-profiles" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.403150 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.406320 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.406514 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.406570 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.406656 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.412677 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-n5fvd"] Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.525857 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ceph\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.525987 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.526245 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-inventory\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.526457 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65n2z\" (UniqueName: \"kubernetes.io/projected/6f58ba99-974b-4ace-bb5a-d01bb707de07-kube-api-access-65n2z\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.628770 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.628942 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-inventory\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.629260 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65n2z\" (UniqueName: \"kubernetes.io/projected/6f58ba99-974b-4ace-bb5a-d01bb707de07-kube-api-access-65n2z\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.629681 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ceph\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.633878 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.635482 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-inventory\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.647658 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ceph\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.648850 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65n2z\" (UniqueName: \"kubernetes.io/projected/6f58ba99-974b-4ace-bb5a-d01bb707de07-kube-api-access-65n2z\") pod \"ceph-client-openstack-openstack-cell1-n5fvd\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:08 crc kubenswrapper[4741]: I0929 21:15:08.726646 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:09 crc kubenswrapper[4741]: I0929 21:15:09.300609 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-n5fvd"] Sep 29 21:15:10 crc kubenswrapper[4741]: I0929 21:15:10.316942 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" event={"ID":"6f58ba99-974b-4ace-bb5a-d01bb707de07","Type":"ContainerStarted","Data":"88ed60f98bd594c07bfe55c7f7cd07ef20b7817cf4aa909dda7d86a45538b290"} Sep 29 21:15:11 crc kubenswrapper[4741]: I0929 21:15:11.326204 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" event={"ID":"6f58ba99-974b-4ace-bb5a-d01bb707de07","Type":"ContainerStarted","Data":"bcab877d1cd3dd5600280d6772b92273b9fc3e688b23c1eca278be0087e38eb2"} Sep 29 21:15:11 crc kubenswrapper[4741]: I0929 21:15:11.345607 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" podStartSLOduration=1.9994675229999999 podStartE2EDuration="3.345584578s" podCreationTimestamp="2025-09-29 21:15:08 +0000 UTC" firstStartedPulling="2025-09-29 21:15:09.303681663 +0000 UTC m=+7550.951471005" lastFinishedPulling="2025-09-29 21:15:10.649798728 +0000 UTC m=+7552.297588060" observedRunningTime="2025-09-29 21:15:11.340250901 +0000 UTC m=+7552.988040233" watchObservedRunningTime="2025-09-29 21:15:11.345584578 +0000 UTC m=+7552.993373910" Sep 29 21:15:16 crc kubenswrapper[4741]: I0929 21:15:16.086052 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:15:16 crc kubenswrapper[4741]: E0929 21:15:16.090763 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:15:16 crc kubenswrapper[4741]: I0929 21:15:16.379754 4741 generic.go:334] "Generic (PLEG): container finished" podID="6f58ba99-974b-4ace-bb5a-d01bb707de07" containerID="bcab877d1cd3dd5600280d6772b92273b9fc3e688b23c1eca278be0087e38eb2" exitCode=0 Sep 29 21:15:16 crc kubenswrapper[4741]: I0929 21:15:16.379793 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" event={"ID":"6f58ba99-974b-4ace-bb5a-d01bb707de07","Type":"ContainerDied","Data":"bcab877d1cd3dd5600280d6772b92273b9fc3e688b23c1eca278be0087e38eb2"} Sep 29 21:15:17 crc kubenswrapper[4741]: I0929 21:15:17.899401 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.049580 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65n2z\" (UniqueName: \"kubernetes.io/projected/6f58ba99-974b-4ace-bb5a-d01bb707de07-kube-api-access-65n2z\") pod \"6f58ba99-974b-4ace-bb5a-d01bb707de07\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.049748 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ssh-key\") pod \"6f58ba99-974b-4ace-bb5a-d01bb707de07\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.049772 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ceph\") pod \"6f58ba99-974b-4ace-bb5a-d01bb707de07\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.049943 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-inventory\") pod \"6f58ba99-974b-4ace-bb5a-d01bb707de07\" (UID: \"6f58ba99-974b-4ace-bb5a-d01bb707de07\") " Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.055284 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f58ba99-974b-4ace-bb5a-d01bb707de07-kube-api-access-65n2z" (OuterVolumeSpecName: "kube-api-access-65n2z") pod "6f58ba99-974b-4ace-bb5a-d01bb707de07" (UID: "6f58ba99-974b-4ace-bb5a-d01bb707de07"). InnerVolumeSpecName "kube-api-access-65n2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.055826 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ceph" (OuterVolumeSpecName: "ceph") pod "6f58ba99-974b-4ace-bb5a-d01bb707de07" (UID: "6f58ba99-974b-4ace-bb5a-d01bb707de07"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.082152 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-inventory" (OuterVolumeSpecName: "inventory") pod "6f58ba99-974b-4ace-bb5a-d01bb707de07" (UID: "6f58ba99-974b-4ace-bb5a-d01bb707de07"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.085143 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6f58ba99-974b-4ace-bb5a-d01bb707de07" (UID: "6f58ba99-974b-4ace-bb5a-d01bb707de07"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.153729 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.153762 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65n2z\" (UniqueName: \"kubernetes.io/projected/6f58ba99-974b-4ace-bb5a-d01bb707de07-kube-api-access-65n2z\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.153773 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.153781 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f58ba99-974b-4ace-bb5a-d01bb707de07-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.399843 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" event={"ID":"6f58ba99-974b-4ace-bb5a-d01bb707de07","Type":"ContainerDied","Data":"88ed60f98bd594c07bfe55c7f7cd07ef20b7817cf4aa909dda7d86a45538b290"} Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.400043 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88ed60f98bd594c07bfe55c7f7cd07ef20b7817cf4aa909dda7d86a45538b290" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.399904 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-n5fvd" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.484344 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8lmzv"] Sep 29 21:15:18 crc kubenswrapper[4741]: E0929 21:15:18.484978 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f58ba99-974b-4ace-bb5a-d01bb707de07" containerName="ceph-client-openstack-openstack-cell1" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.485010 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f58ba99-974b-4ace-bb5a-d01bb707de07" containerName="ceph-client-openstack-openstack-cell1" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.485415 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f58ba99-974b-4ace-bb5a-d01bb707de07" containerName="ceph-client-openstack-openstack-cell1" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.486583 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.488461 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.489041 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.489098 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.489452 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.490306 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.497016 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8lmzv"] Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.558885 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.559120 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xxww\" (UniqueName: \"kubernetes.io/projected/3b400c3e-1f0f-4283-8e0f-b62978f3877d-kube-api-access-4xxww\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.559334 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ssh-key\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.559489 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-inventory\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.559559 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ceph\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.559645 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.660277 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xxww\" (UniqueName: \"kubernetes.io/projected/3b400c3e-1f0f-4283-8e0f-b62978f3877d-kube-api-access-4xxww\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.660656 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ssh-key\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.660694 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-inventory\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.660721 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ceph\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.660752 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.660844 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.661662 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.664528 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ssh-key\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.665025 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-inventory\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.665282 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.665318 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ceph\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.682422 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xxww\" (UniqueName: \"kubernetes.io/projected/3b400c3e-1f0f-4283-8e0f-b62978f3877d-kube-api-access-4xxww\") pod \"ovn-openstack-openstack-cell1-8lmzv\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:18 crc kubenswrapper[4741]: I0929 21:15:18.814251 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:15:19 crc kubenswrapper[4741]: I0929 21:15:19.370170 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8lmzv"] Sep 29 21:15:19 crc kubenswrapper[4741]: I0929 21:15:19.411967 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" event={"ID":"3b400c3e-1f0f-4283-8e0f-b62978f3877d","Type":"ContainerStarted","Data":"856d72de3c43f64be1325266af8206cc3125e61d250731bbe6f47da459846c0b"} Sep 29 21:15:20 crc kubenswrapper[4741]: I0929 21:15:20.440765 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:15:21 crc kubenswrapper[4741]: I0929 21:15:21.435420 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" event={"ID":"3b400c3e-1f0f-4283-8e0f-b62978f3877d","Type":"ContainerStarted","Data":"0074e4ea5a2eca982e61ebcc877ff8c971364967b73817d645e28a3020055597"} Sep 29 21:15:21 crc kubenswrapper[4741]: I0929 21:15:21.461418 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" podStartSLOduration=2.399287492 podStartE2EDuration="3.461376734s" podCreationTimestamp="2025-09-29 21:15:18 +0000 UTC" firstStartedPulling="2025-09-29 21:15:19.37610581 +0000 UTC m=+7561.023895142" lastFinishedPulling="2025-09-29 21:15:20.438195052 +0000 UTC m=+7562.085984384" observedRunningTime="2025-09-29 21:15:21.451402461 +0000 UTC m=+7563.099191804" watchObservedRunningTime="2025-09-29 21:15:21.461376734 +0000 UTC m=+7563.109166066" Sep 29 21:15:30 crc kubenswrapper[4741]: I0929 21:15:30.086455 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:15:30 crc kubenswrapper[4741]: E0929 21:15:30.087099 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:15:32 crc kubenswrapper[4741]: I0929 21:15:32.025132 4741 scope.go:117] "RemoveContainer" containerID="decfd7ffa52e99c7d110c5a4ad978904e36dd3d76f2e1bb01f3adac1eb8d3cd8" Sep 29 21:15:42 crc kubenswrapper[4741]: I0929 21:15:42.087023 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:15:42 crc kubenswrapper[4741]: E0929 21:15:42.088196 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:15:55 crc kubenswrapper[4741]: I0929 21:15:55.085994 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:15:55 crc kubenswrapper[4741]: E0929 21:15:55.086633 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:16:10 crc kubenswrapper[4741]: I0929 21:16:10.086649 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:16:10 crc kubenswrapper[4741]: E0929 21:16:10.087373 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:16:23 crc kubenswrapper[4741]: I0929 21:16:23.086318 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:16:23 crc kubenswrapper[4741]: E0929 21:16:23.087059 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:16:24 crc kubenswrapper[4741]: I0929 21:16:24.102503 4741 generic.go:334] "Generic (PLEG): container finished" podID="3b400c3e-1f0f-4283-8e0f-b62978f3877d" containerID="0074e4ea5a2eca982e61ebcc877ff8c971364967b73817d645e28a3020055597" exitCode=0 Sep 29 21:16:24 crc kubenswrapper[4741]: I0929 21:16:24.102582 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" event={"ID":"3b400c3e-1f0f-4283-8e0f-b62978f3877d","Type":"ContainerDied","Data":"0074e4ea5a2eca982e61ebcc877ff8c971364967b73817d645e28a3020055597"} Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.567243 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.740555 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovncontroller-config-0\") pod \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.740920 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovn-combined-ca-bundle\") pod \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.741060 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-inventory\") pod \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.741117 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ssh-key\") pod \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.741143 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xxww\" (UniqueName: \"kubernetes.io/projected/3b400c3e-1f0f-4283-8e0f-b62978f3877d-kube-api-access-4xxww\") pod \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.741226 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ceph\") pod \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\" (UID: \"3b400c3e-1f0f-4283-8e0f-b62978f3877d\") " Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.747005 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3b400c3e-1f0f-4283-8e0f-b62978f3877d" (UID: "3b400c3e-1f0f-4283-8e0f-b62978f3877d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.749623 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ceph" (OuterVolumeSpecName: "ceph") pod "3b400c3e-1f0f-4283-8e0f-b62978f3877d" (UID: "3b400c3e-1f0f-4283-8e0f-b62978f3877d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.750651 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b400c3e-1f0f-4283-8e0f-b62978f3877d-kube-api-access-4xxww" (OuterVolumeSpecName: "kube-api-access-4xxww") pod "3b400c3e-1f0f-4283-8e0f-b62978f3877d" (UID: "3b400c3e-1f0f-4283-8e0f-b62978f3877d"). InnerVolumeSpecName "kube-api-access-4xxww". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.767391 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "3b400c3e-1f0f-4283-8e0f-b62978f3877d" (UID: "3b400c3e-1f0f-4283-8e0f-b62978f3877d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.770087 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b400c3e-1f0f-4283-8e0f-b62978f3877d" (UID: "3b400c3e-1f0f-4283-8e0f-b62978f3877d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.770309 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-inventory" (OuterVolumeSpecName: "inventory") pod "3b400c3e-1f0f-4283-8e0f-b62978f3877d" (UID: "3b400c3e-1f0f-4283-8e0f-b62978f3877d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.844650 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.844686 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.844696 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.844704 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xxww\" (UniqueName: \"kubernetes.io/projected/3b400c3e-1f0f-4283-8e0f-b62978f3877d-kube-api-access-4xxww\") on node \"crc\" DevicePath \"\"" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.844713 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:16:25 crc kubenswrapper[4741]: I0929 21:16:25.844722 4741 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3b400c3e-1f0f-4283-8e0f-b62978f3877d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.122312 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" event={"ID":"3b400c3e-1f0f-4283-8e0f-b62978f3877d","Type":"ContainerDied","Data":"856d72de3c43f64be1325266af8206cc3125e61d250731bbe6f47da459846c0b"} Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.122350 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="856d72de3c43f64be1325266af8206cc3125e61d250731bbe6f47da459846c0b" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.122367 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8lmzv" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.211895 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-27c4v"] Sep 29 21:16:26 crc kubenswrapper[4741]: E0929 21:16:26.212465 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b400c3e-1f0f-4283-8e0f-b62978f3877d" containerName="ovn-openstack-openstack-cell1" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.212484 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b400c3e-1f0f-4283-8e0f-b62978f3877d" containerName="ovn-openstack-openstack-cell1" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.212842 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b400c3e-1f0f-4283-8e0f-b62978f3877d" containerName="ovn-openstack-openstack-cell1" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.213794 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.215677 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.216583 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.216960 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.217763 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.218674 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.220140 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.249460 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-27c4v"] Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.253622 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.253718 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.253778 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.253854 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.253933 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.254014 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wgvd\" (UniqueName: \"kubernetes.io/projected/50889d55-3f51-469f-9dfb-e08d0fef628d-kube-api-access-4wgvd\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.254112 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.356439 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.356509 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.356548 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.356584 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.356631 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.356673 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.356705 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wgvd\" (UniqueName: \"kubernetes.io/projected/50889d55-3f51-469f-9dfb-e08d0fef628d-kube-api-access-4wgvd\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.360710 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.360760 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.361393 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.361458 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.361759 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.363019 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.383253 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wgvd\" (UniqueName: \"kubernetes.io/projected/50889d55-3f51-469f-9dfb-e08d0fef628d-kube-api-access-4wgvd\") pod \"neutron-metadata-openstack-openstack-cell1-27c4v\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:26 crc kubenswrapper[4741]: I0929 21:16:26.534944 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:16:27 crc kubenswrapper[4741]: I0929 21:16:27.110834 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-27c4v"] Sep 29 21:16:27 crc kubenswrapper[4741]: I0929 21:16:27.135356 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" event={"ID":"50889d55-3f51-469f-9dfb-e08d0fef628d","Type":"ContainerStarted","Data":"0beec91931ed08802b6e89cc2e573a44cfa1c90a26fbdd80a82061cdb11cb6fb"} Sep 29 21:16:28 crc kubenswrapper[4741]: I0929 21:16:28.148423 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" event={"ID":"50889d55-3f51-469f-9dfb-e08d0fef628d","Type":"ContainerStarted","Data":"e9af905f7dd9eded363aae2e1c2136d8cbca7a2087669f872c63f5ecedaa77c9"} Sep 29 21:16:28 crc kubenswrapper[4741]: I0929 21:16:28.174838 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" podStartSLOduration=1.482959563 podStartE2EDuration="2.17481965s" podCreationTimestamp="2025-09-29 21:16:26 +0000 UTC" firstStartedPulling="2025-09-29 21:16:27.109069564 +0000 UTC m=+7628.756858896" lastFinishedPulling="2025-09-29 21:16:27.800929631 +0000 UTC m=+7629.448718983" observedRunningTime="2025-09-29 21:16:28.164755025 +0000 UTC m=+7629.812544367" watchObservedRunningTime="2025-09-29 21:16:28.17481965 +0000 UTC m=+7629.822608982" Sep 29 21:16:34 crc kubenswrapper[4741]: I0929 21:16:34.085993 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:16:34 crc kubenswrapper[4741]: E0929 21:16:34.087575 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:16:49 crc kubenswrapper[4741]: I0929 21:16:49.101373 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:16:49 crc kubenswrapper[4741]: E0929 21:16:49.102074 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:17:03 crc kubenswrapper[4741]: I0929 21:17:03.086104 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:17:03 crc kubenswrapper[4741]: E0929 21:17:03.086843 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:17:16 crc kubenswrapper[4741]: I0929 21:17:16.086470 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:17:16 crc kubenswrapper[4741]: E0929 21:17:16.087175 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:17:19 crc kubenswrapper[4741]: I0929 21:17:19.715379 4741 generic.go:334] "Generic (PLEG): container finished" podID="50889d55-3f51-469f-9dfb-e08d0fef628d" containerID="e9af905f7dd9eded363aae2e1c2136d8cbca7a2087669f872c63f5ecedaa77c9" exitCode=0 Sep 29 21:17:19 crc kubenswrapper[4741]: I0929 21:17:19.715498 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" event={"ID":"50889d55-3f51-469f-9dfb-e08d0fef628d","Type":"ContainerDied","Data":"e9af905f7dd9eded363aae2e1c2136d8cbca7a2087669f872c63f5ecedaa77c9"} Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.193053 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.387460 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ceph\") pod \"50889d55-3f51-469f-9dfb-e08d0fef628d\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.387552 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ssh-key\") pod \"50889d55-3f51-469f-9dfb-e08d0fef628d\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.387690 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-nova-metadata-neutron-config-0\") pod \"50889d55-3f51-469f-9dfb-e08d0fef628d\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.388662 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"50889d55-3f51-469f-9dfb-e08d0fef628d\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.388798 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-inventory\") pod \"50889d55-3f51-469f-9dfb-e08d0fef628d\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.388894 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-metadata-combined-ca-bundle\") pod \"50889d55-3f51-469f-9dfb-e08d0fef628d\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.388985 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wgvd\" (UniqueName: \"kubernetes.io/projected/50889d55-3f51-469f-9dfb-e08d0fef628d-kube-api-access-4wgvd\") pod \"50889d55-3f51-469f-9dfb-e08d0fef628d\" (UID: \"50889d55-3f51-469f-9dfb-e08d0fef628d\") " Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.393871 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ceph" (OuterVolumeSpecName: "ceph") pod "50889d55-3f51-469f-9dfb-e08d0fef628d" (UID: "50889d55-3f51-469f-9dfb-e08d0fef628d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.399509 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50889d55-3f51-469f-9dfb-e08d0fef628d-kube-api-access-4wgvd" (OuterVolumeSpecName: "kube-api-access-4wgvd") pod "50889d55-3f51-469f-9dfb-e08d0fef628d" (UID: "50889d55-3f51-469f-9dfb-e08d0fef628d"). InnerVolumeSpecName "kube-api-access-4wgvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.399890 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "50889d55-3f51-469f-9dfb-e08d0fef628d" (UID: "50889d55-3f51-469f-9dfb-e08d0fef628d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.417285 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "50889d55-3f51-469f-9dfb-e08d0fef628d" (UID: "50889d55-3f51-469f-9dfb-e08d0fef628d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.420124 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "50889d55-3f51-469f-9dfb-e08d0fef628d" (UID: "50889d55-3f51-469f-9dfb-e08d0fef628d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.425755 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "50889d55-3f51-469f-9dfb-e08d0fef628d" (UID: "50889d55-3f51-469f-9dfb-e08d0fef628d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.432905 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-inventory" (OuterVolumeSpecName: "inventory") pod "50889d55-3f51-469f-9dfb-e08d0fef628d" (UID: "50889d55-3f51-469f-9dfb-e08d0fef628d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.492971 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.493010 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.493021 4741 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.493030 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.493040 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.493051 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50889d55-3f51-469f-9dfb-e08d0fef628d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.493060 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wgvd\" (UniqueName: \"kubernetes.io/projected/50889d55-3f51-469f-9dfb-e08d0fef628d-kube-api-access-4wgvd\") on node \"crc\" DevicePath \"\"" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.741160 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" event={"ID":"50889d55-3f51-469f-9dfb-e08d0fef628d","Type":"ContainerDied","Data":"0beec91931ed08802b6e89cc2e573a44cfa1c90a26fbdd80a82061cdb11cb6fb"} Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.741534 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0beec91931ed08802b6e89cc2e573a44cfa1c90a26fbdd80a82061cdb11cb6fb" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.741216 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-27c4v" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.839566 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-7qvfs"] Sep 29 21:17:21 crc kubenswrapper[4741]: E0929 21:17:21.841155 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50889d55-3f51-469f-9dfb-e08d0fef628d" containerName="neutron-metadata-openstack-openstack-cell1" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.841384 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="50889d55-3f51-469f-9dfb-e08d0fef628d" containerName="neutron-metadata-openstack-openstack-cell1" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.842747 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="50889d55-3f51-469f-9dfb-e08d0fef628d" containerName="neutron-metadata-openstack-openstack-cell1" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.844946 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.848152 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.849506 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.849651 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.849787 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.849975 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:17:21 crc kubenswrapper[4741]: I0929 21:17:21.881925 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-7qvfs"] Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.004155 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.004239 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4jq8\" (UniqueName: \"kubernetes.io/projected/94f70bce-ae42-46eb-b72e-3c7d105608d3-kube-api-access-q4jq8\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.004283 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ssh-key\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.004371 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-inventory\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.004452 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.004490 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ceph\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.105913 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-inventory\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.105990 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.106017 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ceph\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.106092 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.106139 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4jq8\" (UniqueName: \"kubernetes.io/projected/94f70bce-ae42-46eb-b72e-3c7d105608d3-kube-api-access-q4jq8\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.106184 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ssh-key\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.109823 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.109862 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ssh-key\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.109975 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-inventory\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.111229 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ceph\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.122543 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.125203 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4jq8\" (UniqueName: \"kubernetes.io/projected/94f70bce-ae42-46eb-b72e-3c7d105608d3-kube-api-access-q4jq8\") pod \"libvirt-openstack-openstack-cell1-7qvfs\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.180008 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.774575 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-7qvfs"] Sep 29 21:17:22 crc kubenswrapper[4741]: I0929 21:17:22.780958 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:17:23 crc kubenswrapper[4741]: I0929 21:17:23.760800 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" event={"ID":"94f70bce-ae42-46eb-b72e-3c7d105608d3","Type":"ContainerStarted","Data":"3dbb0f4ef82530dbd1f850de41ea785b54af1a0c48e079ccb1d312c78385d001"} Sep 29 21:17:23 crc kubenswrapper[4741]: I0929 21:17:23.761682 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" event={"ID":"94f70bce-ae42-46eb-b72e-3c7d105608d3","Type":"ContainerStarted","Data":"efd8944f7ec314e7edef04fcee57a31d8fdb92e55e113517717fa01368afa3eb"} Sep 29 21:17:23 crc kubenswrapper[4741]: I0929 21:17:23.782093 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" podStartSLOduration=2.3388876659999998 podStartE2EDuration="2.782064268s" podCreationTimestamp="2025-09-29 21:17:21 +0000 UTC" firstStartedPulling="2025-09-29 21:17:22.780694439 +0000 UTC m=+7684.428483771" lastFinishedPulling="2025-09-29 21:17:23.223871001 +0000 UTC m=+7684.871660373" observedRunningTime="2025-09-29 21:17:23.777083222 +0000 UTC m=+7685.424872604" watchObservedRunningTime="2025-09-29 21:17:23.782064268 +0000 UTC m=+7685.429853600" Sep 29 21:17:29 crc kubenswrapper[4741]: I0929 21:17:29.091868 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:17:29 crc kubenswrapper[4741]: E0929 21:17:29.092582 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:17:43 crc kubenswrapper[4741]: I0929 21:17:43.086262 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:17:43 crc kubenswrapper[4741]: I0929 21:17:43.977093 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"17c795d8138817a2c28c07cebb79298eee61f7e959f13bcffc60142e93418239"} Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.397137 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jkvlk"] Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.405351 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.417168 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jkvlk"] Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.561793 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5zzt\" (UniqueName: \"kubernetes.io/projected/3048841f-507c-4893-bd37-0ee596d14f4b-kube-api-access-f5zzt\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.561900 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-utilities\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.562200 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-catalog-content\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.665989 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-catalog-content\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.666167 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5zzt\" (UniqueName: \"kubernetes.io/projected/3048841f-507c-4893-bd37-0ee596d14f4b-kube-api-access-f5zzt\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.666256 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-utilities\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.666452 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-catalog-content\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.666734 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-utilities\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.688101 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5zzt\" (UniqueName: \"kubernetes.io/projected/3048841f-507c-4893-bd37-0ee596d14f4b-kube-api-access-f5zzt\") pod \"redhat-operators-jkvlk\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:48 crc kubenswrapper[4741]: I0929 21:18:48.724521 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:49 crc kubenswrapper[4741]: I0929 21:18:49.289840 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jkvlk"] Sep 29 21:18:49 crc kubenswrapper[4741]: I0929 21:18:49.678884 4741 generic.go:334] "Generic (PLEG): container finished" podID="3048841f-507c-4893-bd37-0ee596d14f4b" containerID="42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27" exitCode=0 Sep 29 21:18:49 crc kubenswrapper[4741]: I0929 21:18:49.678990 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkvlk" event={"ID":"3048841f-507c-4893-bd37-0ee596d14f4b","Type":"ContainerDied","Data":"42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27"} Sep 29 21:18:49 crc kubenswrapper[4741]: I0929 21:18:49.679237 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkvlk" event={"ID":"3048841f-507c-4893-bd37-0ee596d14f4b","Type":"ContainerStarted","Data":"e454e0212e488899c014414b6eb9c207360271a48fe96877913ef6d56a3ffca1"} Sep 29 21:18:51 crc kubenswrapper[4741]: I0929 21:18:51.701194 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkvlk" event={"ID":"3048841f-507c-4893-bd37-0ee596d14f4b","Type":"ContainerStarted","Data":"3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81"} Sep 29 21:18:55 crc kubenswrapper[4741]: I0929 21:18:55.753569 4741 generic.go:334] "Generic (PLEG): container finished" podID="3048841f-507c-4893-bd37-0ee596d14f4b" containerID="3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81" exitCode=0 Sep 29 21:18:55 crc kubenswrapper[4741]: I0929 21:18:55.753645 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkvlk" event={"ID":"3048841f-507c-4893-bd37-0ee596d14f4b","Type":"ContainerDied","Data":"3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81"} Sep 29 21:18:56 crc kubenswrapper[4741]: I0929 21:18:56.769379 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkvlk" event={"ID":"3048841f-507c-4893-bd37-0ee596d14f4b","Type":"ContainerStarted","Data":"324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b"} Sep 29 21:18:56 crc kubenswrapper[4741]: I0929 21:18:56.789705 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jkvlk" podStartSLOduration=2.330022643 podStartE2EDuration="8.789688896s" podCreationTimestamp="2025-09-29 21:18:48 +0000 UTC" firstStartedPulling="2025-09-29 21:18:49.680556595 +0000 UTC m=+7771.328345917" lastFinishedPulling="2025-09-29 21:18:56.140222838 +0000 UTC m=+7777.788012170" observedRunningTime="2025-09-29 21:18:56.788341464 +0000 UTC m=+7778.436130796" watchObservedRunningTime="2025-09-29 21:18:56.789688896 +0000 UTC m=+7778.437478218" Sep 29 21:18:58 crc kubenswrapper[4741]: I0929 21:18:58.725955 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:58 crc kubenswrapper[4741]: I0929 21:18:58.726305 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:18:59 crc kubenswrapper[4741]: I0929 21:18:59.779790 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jkvlk" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="registry-server" probeResult="failure" output=< Sep 29 21:18:59 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 21:18:59 crc kubenswrapper[4741]: > Sep 29 21:19:08 crc kubenswrapper[4741]: I0929 21:19:08.793819 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:19:08 crc kubenswrapper[4741]: I0929 21:19:08.854620 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:19:09 crc kubenswrapper[4741]: I0929 21:19:09.040214 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jkvlk"] Sep 29 21:19:09 crc kubenswrapper[4741]: I0929 21:19:09.906896 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jkvlk" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="registry-server" containerID="cri-o://324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b" gracePeriod=2 Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.422213 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.490318 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-catalog-content\") pod \"3048841f-507c-4893-bd37-0ee596d14f4b\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.490617 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5zzt\" (UniqueName: \"kubernetes.io/projected/3048841f-507c-4893-bd37-0ee596d14f4b-kube-api-access-f5zzt\") pod \"3048841f-507c-4893-bd37-0ee596d14f4b\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.490665 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-utilities\") pod \"3048841f-507c-4893-bd37-0ee596d14f4b\" (UID: \"3048841f-507c-4893-bd37-0ee596d14f4b\") " Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.491870 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-utilities" (OuterVolumeSpecName: "utilities") pod "3048841f-507c-4893-bd37-0ee596d14f4b" (UID: "3048841f-507c-4893-bd37-0ee596d14f4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.497105 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3048841f-507c-4893-bd37-0ee596d14f4b-kube-api-access-f5zzt" (OuterVolumeSpecName: "kube-api-access-f5zzt") pod "3048841f-507c-4893-bd37-0ee596d14f4b" (UID: "3048841f-507c-4893-bd37-0ee596d14f4b"). InnerVolumeSpecName "kube-api-access-f5zzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.572149 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3048841f-507c-4893-bd37-0ee596d14f4b" (UID: "3048841f-507c-4893-bd37-0ee596d14f4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.594621 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.594653 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5zzt\" (UniqueName: \"kubernetes.io/projected/3048841f-507c-4893-bd37-0ee596d14f4b-kube-api-access-f5zzt\") on node \"crc\" DevicePath \"\"" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.594665 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3048841f-507c-4893-bd37-0ee596d14f4b-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.921027 4741 generic.go:334] "Generic (PLEG): container finished" podID="3048841f-507c-4893-bd37-0ee596d14f4b" containerID="324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b" exitCode=0 Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.921074 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkvlk" event={"ID":"3048841f-507c-4893-bd37-0ee596d14f4b","Type":"ContainerDied","Data":"324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b"} Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.921099 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkvlk" event={"ID":"3048841f-507c-4893-bd37-0ee596d14f4b","Type":"ContainerDied","Data":"e454e0212e488899c014414b6eb9c207360271a48fe96877913ef6d56a3ffca1"} Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.921116 4741 scope.go:117] "RemoveContainer" containerID="324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.921233 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkvlk" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.950617 4741 scope.go:117] "RemoveContainer" containerID="3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81" Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.954209 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jkvlk"] Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.965826 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jkvlk"] Sep 29 21:19:10 crc kubenswrapper[4741]: I0929 21:19:10.984770 4741 scope.go:117] "RemoveContainer" containerID="42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27" Sep 29 21:19:11 crc kubenswrapper[4741]: I0929 21:19:11.028216 4741 scope.go:117] "RemoveContainer" containerID="324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b" Sep 29 21:19:11 crc kubenswrapper[4741]: E0929 21:19:11.028816 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b\": container with ID starting with 324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b not found: ID does not exist" containerID="324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b" Sep 29 21:19:11 crc kubenswrapper[4741]: I0929 21:19:11.028868 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b"} err="failed to get container status \"324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b\": rpc error: code = NotFound desc = could not find container \"324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b\": container with ID starting with 324fdf65c34220995314638b98036059ecbad5a9342cabe0966b63d6285d948b not found: ID does not exist" Sep 29 21:19:11 crc kubenswrapper[4741]: I0929 21:19:11.028904 4741 scope.go:117] "RemoveContainer" containerID="3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81" Sep 29 21:19:11 crc kubenswrapper[4741]: E0929 21:19:11.029542 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81\": container with ID starting with 3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81 not found: ID does not exist" containerID="3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81" Sep 29 21:19:11 crc kubenswrapper[4741]: I0929 21:19:11.029574 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81"} err="failed to get container status \"3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81\": rpc error: code = NotFound desc = could not find container \"3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81\": container with ID starting with 3698a1f4342bb6f2324b35582d8b0ad0d070e2316cc6182fd156b26275215b81 not found: ID does not exist" Sep 29 21:19:11 crc kubenswrapper[4741]: I0929 21:19:11.029609 4741 scope.go:117] "RemoveContainer" containerID="42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27" Sep 29 21:19:11 crc kubenswrapper[4741]: E0929 21:19:11.029943 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27\": container with ID starting with 42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27 not found: ID does not exist" containerID="42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27" Sep 29 21:19:11 crc kubenswrapper[4741]: I0929 21:19:11.029998 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27"} err="failed to get container status \"42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27\": rpc error: code = NotFound desc = could not find container \"42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27\": container with ID starting with 42fdee1168cc1ea867b2f135ccb33cc51a883bfef4d0c0bb5a4c79655582df27 not found: ID does not exist" Sep 29 21:19:11 crc kubenswrapper[4741]: I0929 21:19:11.097825 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" path="/var/lib/kubelet/pods/3048841f-507c-4893-bd37-0ee596d14f4b/volumes" Sep 29 21:20:01 crc kubenswrapper[4741]: I0929 21:20:01.739382 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:20:01 crc kubenswrapper[4741]: I0929 21:20:01.739970 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:20:31 crc kubenswrapper[4741]: I0929 21:20:31.739644 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:20:31 crc kubenswrapper[4741]: I0929 21:20:31.742030 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:21:01 crc kubenswrapper[4741]: I0929 21:21:01.738593 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:21:01 crc kubenswrapper[4741]: I0929 21:21:01.739172 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:21:01 crc kubenswrapper[4741]: I0929 21:21:01.739209 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:21:01 crc kubenswrapper[4741]: I0929 21:21:01.739876 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17c795d8138817a2c28c07cebb79298eee61f7e959f13bcffc60142e93418239"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:21:01 crc kubenswrapper[4741]: I0929 21:21:01.739929 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://17c795d8138817a2c28c07cebb79298eee61f7e959f13bcffc60142e93418239" gracePeriod=600 Sep 29 21:21:02 crc kubenswrapper[4741]: I0929 21:21:02.150889 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="17c795d8138817a2c28c07cebb79298eee61f7e959f13bcffc60142e93418239" exitCode=0 Sep 29 21:21:02 crc kubenswrapper[4741]: I0929 21:21:02.150950 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"17c795d8138817a2c28c07cebb79298eee61f7e959f13bcffc60142e93418239"} Sep 29 21:21:02 crc kubenswrapper[4741]: I0929 21:21:02.151340 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c"} Sep 29 21:21:02 crc kubenswrapper[4741]: I0929 21:21:02.151504 4741 scope.go:117] "RemoveContainer" containerID="0b16bc5dce390cdc5cf86d0f6a646c39ea98856276a021c4481373169394e609" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.309503 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fhkw8"] Sep 29 21:21:07 crc kubenswrapper[4741]: E0929 21:21:07.310447 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="extract-content" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.310460 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="extract-content" Sep 29 21:21:07 crc kubenswrapper[4741]: E0929 21:21:07.310474 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="registry-server" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.310481 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="registry-server" Sep 29 21:21:07 crc kubenswrapper[4741]: E0929 21:21:07.310526 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="extract-utilities" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.310532 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="extract-utilities" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.310752 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="3048841f-507c-4893-bd37-0ee596d14f4b" containerName="registry-server" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.312318 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.325486 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fhkw8"] Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.462529 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-utilities\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.462893 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-catalog-content\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.462976 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85svp\" (UniqueName: \"kubernetes.io/projected/edb94524-cb7e-4af0-80f7-4671a1f67d06-kube-api-access-85svp\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.564653 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-utilities\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.564759 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-catalog-content\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.564833 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85svp\" (UniqueName: \"kubernetes.io/projected/edb94524-cb7e-4af0-80f7-4671a1f67d06-kube-api-access-85svp\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.565972 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-catalog-content\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.566091 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-utilities\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.597465 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85svp\" (UniqueName: \"kubernetes.io/projected/edb94524-cb7e-4af0-80f7-4671a1f67d06-kube-api-access-85svp\") pod \"certified-operators-fhkw8\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:07 crc kubenswrapper[4741]: I0929 21:21:07.635969 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:08 crc kubenswrapper[4741]: I0929 21:21:08.200462 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fhkw8"] Sep 29 21:21:08 crc kubenswrapper[4741]: I0929 21:21:08.221013 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhkw8" event={"ID":"edb94524-cb7e-4af0-80f7-4671a1f67d06","Type":"ContainerStarted","Data":"534e4ca6b469080cfbdd3ca27f2638196d5235fac8f50b9869f3e365921e3cac"} Sep 29 21:21:09 crc kubenswrapper[4741]: I0929 21:21:09.231361 4741 generic.go:334] "Generic (PLEG): container finished" podID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerID="c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a" exitCode=0 Sep 29 21:21:09 crc kubenswrapper[4741]: I0929 21:21:09.232537 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhkw8" event={"ID":"edb94524-cb7e-4af0-80f7-4671a1f67d06","Type":"ContainerDied","Data":"c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a"} Sep 29 21:21:11 crc kubenswrapper[4741]: I0929 21:21:11.264571 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhkw8" event={"ID":"edb94524-cb7e-4af0-80f7-4671a1f67d06","Type":"ContainerStarted","Data":"100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466"} Sep 29 21:21:12 crc kubenswrapper[4741]: I0929 21:21:12.279741 4741 generic.go:334] "Generic (PLEG): container finished" podID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerID="100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466" exitCode=0 Sep 29 21:21:12 crc kubenswrapper[4741]: I0929 21:21:12.279783 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhkw8" event={"ID":"edb94524-cb7e-4af0-80f7-4671a1f67d06","Type":"ContainerDied","Data":"100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466"} Sep 29 21:21:13 crc kubenswrapper[4741]: I0929 21:21:13.292498 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhkw8" event={"ID":"edb94524-cb7e-4af0-80f7-4671a1f67d06","Type":"ContainerStarted","Data":"93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597"} Sep 29 21:21:13 crc kubenswrapper[4741]: I0929 21:21:13.313978 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fhkw8" podStartSLOduration=2.82166811 podStartE2EDuration="6.313961976s" podCreationTimestamp="2025-09-29 21:21:07 +0000 UTC" firstStartedPulling="2025-09-29 21:21:09.233777969 +0000 UTC m=+7910.881567301" lastFinishedPulling="2025-09-29 21:21:12.726071795 +0000 UTC m=+7914.373861167" observedRunningTime="2025-09-29 21:21:13.311215829 +0000 UTC m=+7914.959005161" watchObservedRunningTime="2025-09-29 21:21:13.313961976 +0000 UTC m=+7914.961751298" Sep 29 21:21:17 crc kubenswrapper[4741]: I0929 21:21:17.637466 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:17 crc kubenswrapper[4741]: I0929 21:21:17.638148 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:17 crc kubenswrapper[4741]: I0929 21:21:17.709434 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:18 crc kubenswrapper[4741]: I0929 21:21:18.382769 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:18 crc kubenswrapper[4741]: I0929 21:21:18.441028 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fhkw8"] Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.362349 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fhkw8" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="registry-server" containerID="cri-o://93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597" gracePeriod=2 Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.374066 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xh6pp"] Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.376571 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.406118 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh6pp"] Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.567008 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t7l9\" (UniqueName: \"kubernetes.io/projected/55344917-b672-42b9-aef3-140b149d8a66-kube-api-access-6t7l9\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.567409 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-catalog-content\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.567486 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-utilities\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: E0929 21:21:20.595851 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedb94524_cb7e_4af0_80f7_4671a1f67d06.slice/crio-93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597.scope\": RecentStats: unable to find data in memory cache]" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.669691 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-catalog-content\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.669975 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-utilities\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.670154 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t7l9\" (UniqueName: \"kubernetes.io/projected/55344917-b672-42b9-aef3-140b149d8a66-kube-api-access-6t7l9\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.671214 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-catalog-content\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.671277 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-utilities\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.687362 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t7l9\" (UniqueName: \"kubernetes.io/projected/55344917-b672-42b9-aef3-140b149d8a66-kube-api-access-6t7l9\") pod \"redhat-marketplace-xh6pp\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.713086 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:20 crc kubenswrapper[4741]: I0929 21:21:20.892454 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.076794 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-utilities\") pod \"edb94524-cb7e-4af0-80f7-4671a1f67d06\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.077358 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-catalog-content\") pod \"edb94524-cb7e-4af0-80f7-4671a1f67d06\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.077431 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85svp\" (UniqueName: \"kubernetes.io/projected/edb94524-cb7e-4af0-80f7-4671a1f67d06-kube-api-access-85svp\") pod \"edb94524-cb7e-4af0-80f7-4671a1f67d06\" (UID: \"edb94524-cb7e-4af0-80f7-4671a1f67d06\") " Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.077996 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-utilities" (OuterVolumeSpecName: "utilities") pod "edb94524-cb7e-4af0-80f7-4671a1f67d06" (UID: "edb94524-cb7e-4af0-80f7-4671a1f67d06"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.082589 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb94524-cb7e-4af0-80f7-4671a1f67d06-kube-api-access-85svp" (OuterVolumeSpecName: "kube-api-access-85svp") pod "edb94524-cb7e-4af0-80f7-4671a1f67d06" (UID: "edb94524-cb7e-4af0-80f7-4671a1f67d06"). InnerVolumeSpecName "kube-api-access-85svp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.180494 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85svp\" (UniqueName: \"kubernetes.io/projected/edb94524-cb7e-4af0-80f7-4671a1f67d06-kube-api-access-85svp\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.180531 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:21 crc kubenswrapper[4741]: W0929 21:21:21.209873 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55344917_b672_42b9_aef3_140b149d8a66.slice/crio-e6fe26b4a0dfe4a418f567b9e19738d5d223cc2c522091c75f18dea87f2e0fae WatchSource:0}: Error finding container e6fe26b4a0dfe4a418f567b9e19738d5d223cc2c522091c75f18dea87f2e0fae: Status 404 returned error can't find the container with id e6fe26b4a0dfe4a418f567b9e19738d5d223cc2c522091c75f18dea87f2e0fae Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.211100 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh6pp"] Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.282529 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edb94524-cb7e-4af0-80f7-4671a1f67d06" (UID: "edb94524-cb7e-4af0-80f7-4671a1f67d06"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.375831 4741 generic.go:334] "Generic (PLEG): container finished" podID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerID="93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597" exitCode=0 Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.375888 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhkw8" event={"ID":"edb94524-cb7e-4af0-80f7-4671a1f67d06","Type":"ContainerDied","Data":"93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597"} Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.375972 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhkw8" event={"ID":"edb94524-cb7e-4af0-80f7-4671a1f67d06","Type":"ContainerDied","Data":"534e4ca6b469080cfbdd3ca27f2638196d5235fac8f50b9869f3e365921e3cac"} Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.375992 4741 scope.go:117] "RemoveContainer" containerID="93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.375850 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhkw8" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.377807 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh6pp" event={"ID":"55344917-b672-42b9-aef3-140b149d8a66","Type":"ContainerStarted","Data":"e6fe26b4a0dfe4a418f567b9e19738d5d223cc2c522091c75f18dea87f2e0fae"} Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.385744 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb94524-cb7e-4af0-80f7-4671a1f67d06-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.396969 4741 scope.go:117] "RemoveContainer" containerID="100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.493934 4741 scope.go:117] "RemoveContainer" containerID="c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.500527 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fhkw8"] Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.509669 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fhkw8"] Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.578304 4741 scope.go:117] "RemoveContainer" containerID="93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597" Sep 29 21:21:21 crc kubenswrapper[4741]: E0929 21:21:21.578833 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597\": container with ID starting with 93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597 not found: ID does not exist" containerID="93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.578865 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597"} err="failed to get container status \"93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597\": rpc error: code = NotFound desc = could not find container \"93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597\": container with ID starting with 93d03ac753f90a355b7230b37bf2fcd4a776f38df5c659d1df33814343f5c597 not found: ID does not exist" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.578886 4741 scope.go:117] "RemoveContainer" containerID="100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466" Sep 29 21:21:21 crc kubenswrapper[4741]: E0929 21:21:21.579472 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466\": container with ID starting with 100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466 not found: ID does not exist" containerID="100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.579497 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466"} err="failed to get container status \"100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466\": rpc error: code = NotFound desc = could not find container \"100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466\": container with ID starting with 100f7b87728f9177182012ddbc4bb8a64fcc45a52dd582db11149a40d6421466 not found: ID does not exist" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.579512 4741 scope.go:117] "RemoveContainer" containerID="c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a" Sep 29 21:21:21 crc kubenswrapper[4741]: E0929 21:21:21.579793 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a\": container with ID starting with c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a not found: ID does not exist" containerID="c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a" Sep 29 21:21:21 crc kubenswrapper[4741]: I0929 21:21:21.579875 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a"} err="failed to get container status \"c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a\": rpc error: code = NotFound desc = could not find container \"c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a\": container with ID starting with c8982f32bc2c546c2ede623b4e1a3ec6deca255b88ecb692f064f0f2f1b7685a not found: ID does not exist" Sep 29 21:21:22 crc kubenswrapper[4741]: I0929 21:21:22.401185 4741 generic.go:334] "Generic (PLEG): container finished" podID="55344917-b672-42b9-aef3-140b149d8a66" containerID="997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82" exitCode=0 Sep 29 21:21:22 crc kubenswrapper[4741]: I0929 21:21:22.401434 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh6pp" event={"ID":"55344917-b672-42b9-aef3-140b149d8a66","Type":"ContainerDied","Data":"997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82"} Sep 29 21:21:23 crc kubenswrapper[4741]: I0929 21:21:23.103293 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" path="/var/lib/kubelet/pods/edb94524-cb7e-4af0-80f7-4671a1f67d06/volumes" Sep 29 21:21:24 crc kubenswrapper[4741]: I0929 21:21:24.449372 4741 generic.go:334] "Generic (PLEG): container finished" podID="55344917-b672-42b9-aef3-140b149d8a66" containerID="d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a" exitCode=0 Sep 29 21:21:24 crc kubenswrapper[4741]: I0929 21:21:24.450223 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh6pp" event={"ID":"55344917-b672-42b9-aef3-140b149d8a66","Type":"ContainerDied","Data":"d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a"} Sep 29 21:21:25 crc kubenswrapper[4741]: I0929 21:21:25.466689 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh6pp" event={"ID":"55344917-b672-42b9-aef3-140b149d8a66","Type":"ContainerStarted","Data":"4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1"} Sep 29 21:21:25 crc kubenswrapper[4741]: I0929 21:21:25.484350 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xh6pp" podStartSLOduration=3.002645961 podStartE2EDuration="5.48432929s" podCreationTimestamp="2025-09-29 21:21:20 +0000 UTC" firstStartedPulling="2025-09-29 21:21:22.403474853 +0000 UTC m=+7924.051264185" lastFinishedPulling="2025-09-29 21:21:24.885158182 +0000 UTC m=+7926.532947514" observedRunningTime="2025-09-29 21:21:25.482344467 +0000 UTC m=+7927.130133799" watchObservedRunningTime="2025-09-29 21:21:25.48432929 +0000 UTC m=+7927.132118622" Sep 29 21:21:30 crc kubenswrapper[4741]: I0929 21:21:30.715337 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:30 crc kubenswrapper[4741]: I0929 21:21:30.716051 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:30 crc kubenswrapper[4741]: I0929 21:21:30.806925 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:31 crc kubenswrapper[4741]: I0929 21:21:31.630088 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:31 crc kubenswrapper[4741]: I0929 21:21:31.687602 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh6pp"] Sep 29 21:21:33 crc kubenswrapper[4741]: I0929 21:21:33.576863 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xh6pp" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="registry-server" containerID="cri-o://4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1" gracePeriod=2 Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.034839 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.198536 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-utilities\") pod \"55344917-b672-42b9-aef3-140b149d8a66\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.198617 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t7l9\" (UniqueName: \"kubernetes.io/projected/55344917-b672-42b9-aef3-140b149d8a66-kube-api-access-6t7l9\") pod \"55344917-b672-42b9-aef3-140b149d8a66\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.198890 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-catalog-content\") pod \"55344917-b672-42b9-aef3-140b149d8a66\" (UID: \"55344917-b672-42b9-aef3-140b149d8a66\") " Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.200493 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-utilities" (OuterVolumeSpecName: "utilities") pod "55344917-b672-42b9-aef3-140b149d8a66" (UID: "55344917-b672-42b9-aef3-140b149d8a66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.209755 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55344917-b672-42b9-aef3-140b149d8a66-kube-api-access-6t7l9" (OuterVolumeSpecName: "kube-api-access-6t7l9") pod "55344917-b672-42b9-aef3-140b149d8a66" (UID: "55344917-b672-42b9-aef3-140b149d8a66"). InnerVolumeSpecName "kube-api-access-6t7l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.223975 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55344917-b672-42b9-aef3-140b149d8a66" (UID: "55344917-b672-42b9-aef3-140b149d8a66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.302033 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.302091 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t7l9\" (UniqueName: \"kubernetes.io/projected/55344917-b672-42b9-aef3-140b149d8a66-kube-api-access-6t7l9\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.302105 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55344917-b672-42b9-aef3-140b149d8a66-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.591911 4741 generic.go:334] "Generic (PLEG): container finished" podID="55344917-b672-42b9-aef3-140b149d8a66" containerID="4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1" exitCode=0 Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.591966 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh6pp" event={"ID":"55344917-b672-42b9-aef3-140b149d8a66","Type":"ContainerDied","Data":"4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1"} Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.592006 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh6pp" event={"ID":"55344917-b672-42b9-aef3-140b149d8a66","Type":"ContainerDied","Data":"e6fe26b4a0dfe4a418f567b9e19738d5d223cc2c522091c75f18dea87f2e0fae"} Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.592026 4741 scope.go:117] "RemoveContainer" containerID="4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.592076 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh6pp" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.629288 4741 scope.go:117] "RemoveContainer" containerID="d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.637805 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh6pp"] Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.646902 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh6pp"] Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.681426 4741 scope.go:117] "RemoveContainer" containerID="997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.726559 4741 scope.go:117] "RemoveContainer" containerID="4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1" Sep 29 21:21:34 crc kubenswrapper[4741]: E0929 21:21:34.727166 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1\": container with ID starting with 4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1 not found: ID does not exist" containerID="4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.727218 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1"} err="failed to get container status \"4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1\": rpc error: code = NotFound desc = could not find container \"4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1\": container with ID starting with 4e545ed4c6cb66d86809e4afb2790dd8fa9ee0252a8aca8779be438820f0b0f1 not found: ID does not exist" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.727256 4741 scope.go:117] "RemoveContainer" containerID="d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a" Sep 29 21:21:34 crc kubenswrapper[4741]: E0929 21:21:34.727796 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a\": container with ID starting with d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a not found: ID does not exist" containerID="d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.727842 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a"} err="failed to get container status \"d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a\": rpc error: code = NotFound desc = could not find container \"d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a\": container with ID starting with d218a0e590637b940c43fee005fa8db615b50e83b3f1bd1b56704b58a0bbe82a not found: ID does not exist" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.727878 4741 scope.go:117] "RemoveContainer" containerID="997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82" Sep 29 21:21:34 crc kubenswrapper[4741]: E0929 21:21:34.728229 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82\": container with ID starting with 997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82 not found: ID does not exist" containerID="997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82" Sep 29 21:21:34 crc kubenswrapper[4741]: I0929 21:21:34.728288 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82"} err="failed to get container status \"997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82\": rpc error: code = NotFound desc = could not find container \"997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82\": container with ID starting with 997b8803e676df487e5fcf24097c166e1f7d7b1f0761df321846972c18a68a82 not found: ID does not exist" Sep 29 21:21:35 crc kubenswrapper[4741]: I0929 21:21:35.106561 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55344917-b672-42b9-aef3-140b149d8a66" path="/var/lib/kubelet/pods/55344917-b672-42b9-aef3-140b149d8a66/volumes" Sep 29 21:21:51 crc kubenswrapper[4741]: I0929 21:21:51.823574 4741 generic.go:334] "Generic (PLEG): container finished" podID="94f70bce-ae42-46eb-b72e-3c7d105608d3" containerID="3dbb0f4ef82530dbd1f850de41ea785b54af1a0c48e079ccb1d312c78385d001" exitCode=0 Sep 29 21:21:51 crc kubenswrapper[4741]: I0929 21:21:51.823688 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" event={"ID":"94f70bce-ae42-46eb-b72e-3c7d105608d3","Type":"ContainerDied","Data":"3dbb0f4ef82530dbd1f850de41ea785b54af1a0c48e079ccb1d312c78385d001"} Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.370632 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.475293 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ceph\") pod \"94f70bce-ae42-46eb-b72e-3c7d105608d3\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.475350 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-inventory\") pod \"94f70bce-ae42-46eb-b72e-3c7d105608d3\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.475458 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-secret-0\") pod \"94f70bce-ae42-46eb-b72e-3c7d105608d3\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.475526 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ssh-key\") pod \"94f70bce-ae42-46eb-b72e-3c7d105608d3\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.475604 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4jq8\" (UniqueName: \"kubernetes.io/projected/94f70bce-ae42-46eb-b72e-3c7d105608d3-kube-api-access-q4jq8\") pod \"94f70bce-ae42-46eb-b72e-3c7d105608d3\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.475708 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-combined-ca-bundle\") pod \"94f70bce-ae42-46eb-b72e-3c7d105608d3\" (UID: \"94f70bce-ae42-46eb-b72e-3c7d105608d3\") " Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.507561 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "94f70bce-ae42-46eb-b72e-3c7d105608d3" (UID: "94f70bce-ae42-46eb-b72e-3c7d105608d3"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.511640 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ceph" (OuterVolumeSpecName: "ceph") pod "94f70bce-ae42-46eb-b72e-3c7d105608d3" (UID: "94f70bce-ae42-46eb-b72e-3c7d105608d3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.568749 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f70bce-ae42-46eb-b72e-3c7d105608d3-kube-api-access-q4jq8" (OuterVolumeSpecName: "kube-api-access-q4jq8") pod "94f70bce-ae42-46eb-b72e-3c7d105608d3" (UID: "94f70bce-ae42-46eb-b72e-3c7d105608d3"). InnerVolumeSpecName "kube-api-access-q4jq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.568882 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "94f70bce-ae42-46eb-b72e-3c7d105608d3" (UID: "94f70bce-ae42-46eb-b72e-3c7d105608d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.578272 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.578312 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4jq8\" (UniqueName: \"kubernetes.io/projected/94f70bce-ae42-46eb-b72e-3c7d105608d3-kube-api-access-q4jq8\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.578325 4741 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.578336 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.619589 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "94f70bce-ae42-46eb-b72e-3c7d105608d3" (UID: "94f70bce-ae42-46eb-b72e-3c7d105608d3"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.643686 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-inventory" (OuterVolumeSpecName: "inventory") pod "94f70bce-ae42-46eb-b72e-3c7d105608d3" (UID: "94f70bce-ae42-46eb-b72e-3c7d105608d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.680105 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.680137 4741 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94f70bce-ae42-46eb-b72e-3c7d105608d3-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.845952 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" event={"ID":"94f70bce-ae42-46eb-b72e-3c7d105608d3","Type":"ContainerDied","Data":"efd8944f7ec314e7edef04fcee57a31d8fdb92e55e113517717fa01368afa3eb"} Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.845990 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efd8944f7ec314e7edef04fcee57a31d8fdb92e55e113517717fa01368afa3eb" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.846044 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7qvfs" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.926287 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qqgg6"] Sep 29 21:21:53 crc kubenswrapper[4741]: E0929 21:21:53.926957 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f70bce-ae42-46eb-b72e-3c7d105608d3" containerName="libvirt-openstack-openstack-cell1" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.926980 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f70bce-ae42-46eb-b72e-3c7d105608d3" containerName="libvirt-openstack-openstack-cell1" Sep 29 21:21:53 crc kubenswrapper[4741]: E0929 21:21:53.927010 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="extract-content" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927019 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="extract-content" Sep 29 21:21:53 crc kubenswrapper[4741]: E0929 21:21:53.927040 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="registry-server" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927050 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="registry-server" Sep 29 21:21:53 crc kubenswrapper[4741]: E0929 21:21:53.927064 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="extract-utilities" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927072 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="extract-utilities" Sep 29 21:21:53 crc kubenswrapper[4741]: E0929 21:21:53.927092 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="extract-content" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927098 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="extract-content" Sep 29 21:21:53 crc kubenswrapper[4741]: E0929 21:21:53.927133 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="extract-utilities" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927141 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="extract-utilities" Sep 29 21:21:53 crc kubenswrapper[4741]: E0929 21:21:53.927149 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="registry-server" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927156 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="registry-server" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927417 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="55344917-b672-42b9-aef3-140b149d8a66" containerName="registry-server" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927453 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f70bce-ae42-46eb-b72e-3c7d105608d3" containerName="libvirt-openstack-openstack-cell1" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.927475 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb94524-cb7e-4af0-80f7-4671a1f67d06" containerName="registry-server" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.928543 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.930321 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.930780 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.931028 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.931245 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.931611 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.934325 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.934584 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qqgg6"] Sep 29 21:21:53 crc kubenswrapper[4741]: I0929 21:21:53.934627 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.086960 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087318 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087364 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087384 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087477 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxx26\" (UniqueName: \"kubernetes.io/projected/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-kube-api-access-dxx26\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087500 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087564 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087638 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087663 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087748 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.087828 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189484 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189596 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189661 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189789 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189811 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189850 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189867 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189903 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxx26\" (UniqueName: \"kubernetes.io/projected/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-kube-api-access-dxx26\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189920 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.189961 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.190002 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.190756 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.192526 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.195374 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.195414 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.195508 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.196535 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.196795 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.197039 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.197068 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.197784 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.210864 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxx26\" (UniqueName: \"kubernetes.io/projected/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-kube-api-access-dxx26\") pod \"nova-cell1-openstack-openstack-cell1-qqgg6\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.252765 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.814057 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qqgg6"] Sep 29 21:21:54 crc kubenswrapper[4741]: I0929 21:21:54.854981 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" event={"ID":"2f4bee73-3b5c-46f4-9ab7-1631dc495df3","Type":"ContainerStarted","Data":"2dfd9c60524380237274263ee2033b46f236653789cb6c6f269991cfddd76e9c"} Sep 29 21:21:55 crc kubenswrapper[4741]: I0929 21:21:55.869118 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" event={"ID":"2f4bee73-3b5c-46f4-9ab7-1631dc495df3","Type":"ContainerStarted","Data":"158490ba8b64b700612235c38102fa1ca7ee8523c5dc37caad6584eaf2af14a3"} Sep 29 21:21:55 crc kubenswrapper[4741]: I0929 21:21:55.894034 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" podStartSLOduration=2.375685433 podStartE2EDuration="2.894015094s" podCreationTimestamp="2025-09-29 21:21:53 +0000 UTC" firstStartedPulling="2025-09-29 21:21:54.821616181 +0000 UTC m=+7956.469405513" lastFinishedPulling="2025-09-29 21:21:55.339945842 +0000 UTC m=+7956.987735174" observedRunningTime="2025-09-29 21:21:55.885184925 +0000 UTC m=+7957.532974287" watchObservedRunningTime="2025-09-29 21:21:55.894015094 +0000 UTC m=+7957.541804426" Sep 29 21:23:31 crc kubenswrapper[4741]: I0929 21:23:31.739963 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:23:31 crc kubenswrapper[4741]: I0929 21:23:31.740920 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.679020 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5zfxj"] Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.682488 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.692414 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zfxj"] Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.726528 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-catalog-content\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.726900 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75vqs\" (UniqueName: \"kubernetes.io/projected/9fbc54be-5be7-449a-bd16-288b7624d564-kube-api-access-75vqs\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.727112 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-utilities\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.829244 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75vqs\" (UniqueName: \"kubernetes.io/projected/9fbc54be-5be7-449a-bd16-288b7624d564-kube-api-access-75vqs\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.829433 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-utilities\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.829479 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-catalog-content\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.829998 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-utilities\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.830016 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-catalog-content\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:37 crc kubenswrapper[4741]: I0929 21:23:37.850056 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75vqs\" (UniqueName: \"kubernetes.io/projected/9fbc54be-5be7-449a-bd16-288b7624d564-kube-api-access-75vqs\") pod \"community-operators-5zfxj\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:38 crc kubenswrapper[4741]: I0929 21:23:38.003941 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:38 crc kubenswrapper[4741]: I0929 21:23:38.605276 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zfxj"] Sep 29 21:23:38 crc kubenswrapper[4741]: I0929 21:23:38.964803 4741 generic.go:334] "Generic (PLEG): container finished" podID="9fbc54be-5be7-449a-bd16-288b7624d564" containerID="3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f" exitCode=0 Sep 29 21:23:38 crc kubenswrapper[4741]: I0929 21:23:38.964862 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zfxj" event={"ID":"9fbc54be-5be7-449a-bd16-288b7624d564","Type":"ContainerDied","Data":"3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f"} Sep 29 21:23:38 crc kubenswrapper[4741]: I0929 21:23:38.965310 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zfxj" event={"ID":"9fbc54be-5be7-449a-bd16-288b7624d564","Type":"ContainerStarted","Data":"ae6a370fec77abb589ea67d755798e39e2d9b9b7455c77bd11fba963061d60f8"} Sep 29 21:23:38 crc kubenswrapper[4741]: I0929 21:23:38.966904 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:23:39 crc kubenswrapper[4741]: I0929 21:23:39.977692 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zfxj" event={"ID":"9fbc54be-5be7-449a-bd16-288b7624d564","Type":"ContainerStarted","Data":"30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56"} Sep 29 21:23:40 crc kubenswrapper[4741]: I0929 21:23:40.990539 4741 generic.go:334] "Generic (PLEG): container finished" podID="9fbc54be-5be7-449a-bd16-288b7624d564" containerID="30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56" exitCode=0 Sep 29 21:23:40 crc kubenswrapper[4741]: I0929 21:23:40.990862 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zfxj" event={"ID":"9fbc54be-5be7-449a-bd16-288b7624d564","Type":"ContainerDied","Data":"30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56"} Sep 29 21:23:43 crc kubenswrapper[4741]: I0929 21:23:43.014254 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zfxj" event={"ID":"9fbc54be-5be7-449a-bd16-288b7624d564","Type":"ContainerStarted","Data":"d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266"} Sep 29 21:23:43 crc kubenswrapper[4741]: I0929 21:23:43.036875 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5zfxj" podStartSLOduration=3.390465824 podStartE2EDuration="6.036851539s" podCreationTimestamp="2025-09-29 21:23:37 +0000 UTC" firstStartedPulling="2025-09-29 21:23:38.966692749 +0000 UTC m=+8060.614482081" lastFinishedPulling="2025-09-29 21:23:41.613078434 +0000 UTC m=+8063.260867796" observedRunningTime="2025-09-29 21:23:43.030506318 +0000 UTC m=+8064.678295650" watchObservedRunningTime="2025-09-29 21:23:43.036851539 +0000 UTC m=+8064.684640871" Sep 29 21:23:48 crc kubenswrapper[4741]: I0929 21:23:48.004851 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:48 crc kubenswrapper[4741]: I0929 21:23:48.005644 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:48 crc kubenswrapper[4741]: I0929 21:23:48.067955 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:48 crc kubenswrapper[4741]: I0929 21:23:48.136810 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:48 crc kubenswrapper[4741]: I0929 21:23:48.315430 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zfxj"] Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.086220 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5zfxj" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="registry-server" containerID="cri-o://d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266" gracePeriod=2 Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.668475 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.838871 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75vqs\" (UniqueName: \"kubernetes.io/projected/9fbc54be-5be7-449a-bd16-288b7624d564-kube-api-access-75vqs\") pod \"9fbc54be-5be7-449a-bd16-288b7624d564\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.838992 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-catalog-content\") pod \"9fbc54be-5be7-449a-bd16-288b7624d564\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.839147 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-utilities\") pod \"9fbc54be-5be7-449a-bd16-288b7624d564\" (UID: \"9fbc54be-5be7-449a-bd16-288b7624d564\") " Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.840018 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-utilities" (OuterVolumeSpecName: "utilities") pod "9fbc54be-5be7-449a-bd16-288b7624d564" (UID: "9fbc54be-5be7-449a-bd16-288b7624d564"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.843689 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fbc54be-5be7-449a-bd16-288b7624d564-kube-api-access-75vqs" (OuterVolumeSpecName: "kube-api-access-75vqs") pod "9fbc54be-5be7-449a-bd16-288b7624d564" (UID: "9fbc54be-5be7-449a-bd16-288b7624d564"). InnerVolumeSpecName "kube-api-access-75vqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.894093 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fbc54be-5be7-449a-bd16-288b7624d564" (UID: "9fbc54be-5be7-449a-bd16-288b7624d564"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.942283 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75vqs\" (UniqueName: \"kubernetes.io/projected/9fbc54be-5be7-449a-bd16-288b7624d564-kube-api-access-75vqs\") on node \"crc\" DevicePath \"\"" Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.942336 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:23:50 crc kubenswrapper[4741]: I0929 21:23:50.942348 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fbc54be-5be7-449a-bd16-288b7624d564-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.096504 4741 generic.go:334] "Generic (PLEG): container finished" podID="9fbc54be-5be7-449a-bd16-288b7624d564" containerID="d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266" exitCode=0 Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.096594 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zfxj" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.103818 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zfxj" event={"ID":"9fbc54be-5be7-449a-bd16-288b7624d564","Type":"ContainerDied","Data":"d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266"} Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.103869 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zfxj" event={"ID":"9fbc54be-5be7-449a-bd16-288b7624d564","Type":"ContainerDied","Data":"ae6a370fec77abb589ea67d755798e39e2d9b9b7455c77bd11fba963061d60f8"} Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.103892 4741 scope.go:117] "RemoveContainer" containerID="d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.141451 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zfxj"] Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.150004 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5zfxj"] Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.159115 4741 scope.go:117] "RemoveContainer" containerID="30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.210060 4741 scope.go:117] "RemoveContainer" containerID="3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.262745 4741 scope.go:117] "RemoveContainer" containerID="d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266" Sep 29 21:23:51 crc kubenswrapper[4741]: E0929 21:23:51.264862 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266\": container with ID starting with d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266 not found: ID does not exist" containerID="d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.264896 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266"} err="failed to get container status \"d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266\": rpc error: code = NotFound desc = could not find container \"d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266\": container with ID starting with d279b279c9beedc181b17113ff366b3a4c48ef9511fe323e06609b3b85e03266 not found: ID does not exist" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.264919 4741 scope.go:117] "RemoveContainer" containerID="30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56" Sep 29 21:23:51 crc kubenswrapper[4741]: E0929 21:23:51.268751 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56\": container with ID starting with 30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56 not found: ID does not exist" containerID="30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.268795 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56"} err="failed to get container status \"30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56\": rpc error: code = NotFound desc = could not find container \"30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56\": container with ID starting with 30ec676fbdeb3327213dd00b0f43923a42c8bda1a4762709dc7c06392abb7d56 not found: ID does not exist" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.268828 4741 scope.go:117] "RemoveContainer" containerID="3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f" Sep 29 21:23:51 crc kubenswrapper[4741]: E0929 21:23:51.272746 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f\": container with ID starting with 3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f not found: ID does not exist" containerID="3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f" Sep 29 21:23:51 crc kubenswrapper[4741]: I0929 21:23:51.272793 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f"} err="failed to get container status \"3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f\": rpc error: code = NotFound desc = could not find container \"3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f\": container with ID starting with 3a31c74f92fbe6940806ce68f3b7d8dd8fe2052dad736511522797457613826f not found: ID does not exist" Sep 29 21:23:53 crc kubenswrapper[4741]: I0929 21:23:53.097225 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" path="/var/lib/kubelet/pods/9fbc54be-5be7-449a-bd16-288b7624d564/volumes" Sep 29 21:24:01 crc kubenswrapper[4741]: I0929 21:24:01.738765 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:24:01 crc kubenswrapper[4741]: I0929 21:24:01.740769 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:24:31 crc kubenswrapper[4741]: I0929 21:24:31.739469 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:24:31 crc kubenswrapper[4741]: I0929 21:24:31.740015 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:24:31 crc kubenswrapper[4741]: I0929 21:24:31.740057 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:24:31 crc kubenswrapper[4741]: I0929 21:24:31.740857 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:24:31 crc kubenswrapper[4741]: I0929 21:24:31.740913 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" gracePeriod=600 Sep 29 21:24:31 crc kubenswrapper[4741]: E0929 21:24:31.867773 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:24:32 crc kubenswrapper[4741]: I0929 21:24:32.497275 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" exitCode=0 Sep 29 21:24:32 crc kubenswrapper[4741]: I0929 21:24:32.497334 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c"} Sep 29 21:24:32 crc kubenswrapper[4741]: I0929 21:24:32.497385 4741 scope.go:117] "RemoveContainer" containerID="17c795d8138817a2c28c07cebb79298eee61f7e959f13bcffc60142e93418239" Sep 29 21:24:32 crc kubenswrapper[4741]: I0929 21:24:32.498314 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:24:32 crc kubenswrapper[4741]: E0929 21:24:32.498804 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:24:43 crc kubenswrapper[4741]: I0929 21:24:43.085998 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:24:43 crc kubenswrapper[4741]: E0929 21:24:43.087034 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:24:54 crc kubenswrapper[4741]: I0929 21:24:54.086380 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:24:54 crc kubenswrapper[4741]: E0929 21:24:54.087253 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:25:08 crc kubenswrapper[4741]: I0929 21:25:08.086114 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:25:08 crc kubenswrapper[4741]: E0929 21:25:08.087100 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:25:19 crc kubenswrapper[4741]: I0929 21:25:19.093282 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:25:19 crc kubenswrapper[4741]: E0929 21:25:19.094070 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:25:31 crc kubenswrapper[4741]: I0929 21:25:31.086360 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:25:31 crc kubenswrapper[4741]: E0929 21:25:31.087208 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:25:34 crc kubenswrapper[4741]: I0929 21:25:34.091357 4741 generic.go:334] "Generic (PLEG): container finished" podID="2f4bee73-3b5c-46f4-9ab7-1631dc495df3" containerID="158490ba8b64b700612235c38102fa1ca7ee8523c5dc37caad6584eaf2af14a3" exitCode=0 Sep 29 21:25:34 crc kubenswrapper[4741]: I0929 21:25:34.091450 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" event={"ID":"2f4bee73-3b5c-46f4-9ab7-1631dc495df3","Type":"ContainerDied","Data":"158490ba8b64b700612235c38102fa1ca7ee8523c5dc37caad6584eaf2af14a3"} Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.523565 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696148 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-1\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696530 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-0\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696581 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-0\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696618 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ceph\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696642 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ssh-key\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696667 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-1\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696785 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-1\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696818 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-0\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696901 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-inventory\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696934 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxx26\" (UniqueName: \"kubernetes.io/projected/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-kube-api-access-dxx26\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.696956 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-combined-ca-bundle\") pod \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\" (UID: \"2f4bee73-3b5c-46f4-9ab7-1631dc495df3\") " Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.702415 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-kube-api-access-dxx26" (OuterVolumeSpecName: "kube-api-access-dxx26") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "kube-api-access-dxx26". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.702940 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.704601 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ceph" (OuterVolumeSpecName: "ceph") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.738169 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.743110 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.745881 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-inventory" (OuterVolumeSpecName: "inventory") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.751338 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.756260 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.757080 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.759139 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.759639 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2f4bee73-3b5c-46f4-9ab7-1631dc495df3" (UID: "2f4bee73-3b5c-46f4-9ab7-1631dc495df3"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801017 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801053 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxx26\" (UniqueName: \"kubernetes.io/projected/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-kube-api-access-dxx26\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801066 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801076 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801087 4741 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801096 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801129 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801137 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801145 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801155 4741 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:35 crc kubenswrapper[4741]: I0929 21:25:35.801163 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4bee73-3b5c-46f4-9ab7-1631dc495df3-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.111075 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" event={"ID":"2f4bee73-3b5c-46f4-9ab7-1631dc495df3","Type":"ContainerDied","Data":"2dfd9c60524380237274263ee2033b46f236653789cb6c6f269991cfddd76e9c"} Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.111127 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dfd9c60524380237274263ee2033b46f236653789cb6c6f269991cfddd76e9c" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.111138 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qqgg6" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.204825 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-28fg5"] Sep 29 21:25:36 crc kubenswrapper[4741]: E0929 21:25:36.205380 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="registry-server" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.205420 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="registry-server" Sep 29 21:25:36 crc kubenswrapper[4741]: E0929 21:25:36.205462 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="extract-utilities" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.205470 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="extract-utilities" Sep 29 21:25:36 crc kubenswrapper[4741]: E0929 21:25:36.205497 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="extract-content" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.205505 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="extract-content" Sep 29 21:25:36 crc kubenswrapper[4741]: E0929 21:25:36.205519 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4bee73-3b5c-46f4-9ab7-1631dc495df3" containerName="nova-cell1-openstack-openstack-cell1" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.205527 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4bee73-3b5c-46f4-9ab7-1631dc495df3" containerName="nova-cell1-openstack-openstack-cell1" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.205798 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fbc54be-5be7-449a-bd16-288b7624d564" containerName="registry-server" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.205830 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4bee73-3b5c-46f4-9ab7-1631dc495df3" containerName="nova-cell1-openstack-openstack-cell1" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.206789 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.208626 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.209075 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.209082 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.209600 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.213190 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.215961 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-28fg5"] Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.309452 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceph\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.309489 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.309725 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.309911 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.310015 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.310193 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.310427 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvgxq\" (UniqueName: \"kubernetes.io/projected/02f07af8-50b9-4252-90b9-a04657a5916d-kube-api-access-dvgxq\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.310619 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-inventory\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412407 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412500 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvgxq\" (UniqueName: \"kubernetes.io/projected/02f07af8-50b9-4252-90b9-a04657a5916d-kube-api-access-dvgxq\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412568 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-inventory\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412605 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceph\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412621 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412642 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412685 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.412714 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.417312 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.417619 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.417796 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.418282 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.418767 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.419059 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceph\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.419621 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-inventory\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.432312 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvgxq\" (UniqueName: \"kubernetes.io/projected/02f07af8-50b9-4252-90b9-a04657a5916d-kube-api-access-dvgxq\") pod \"telemetry-openstack-openstack-cell1-28fg5\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:36 crc kubenswrapper[4741]: I0929 21:25:36.526066 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:25:37 crc kubenswrapper[4741]: I0929 21:25:37.075749 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-28fg5"] Sep 29 21:25:37 crc kubenswrapper[4741]: I0929 21:25:37.124823 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" event={"ID":"02f07af8-50b9-4252-90b9-a04657a5916d","Type":"ContainerStarted","Data":"edd3eba569d1f58085684d47475aa103080ea0dd449690eed4afbd841c126118"} Sep 29 21:25:38 crc kubenswrapper[4741]: I0929 21:25:38.135801 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" event={"ID":"02f07af8-50b9-4252-90b9-a04657a5916d","Type":"ContainerStarted","Data":"6d4d2acd5bb12f72daf81add5ac54df6bfc49e19049a7c1dcbefafcd180bc8a9"} Sep 29 21:25:38 crc kubenswrapper[4741]: I0929 21:25:38.155684 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" podStartSLOduration=1.618018695 podStartE2EDuration="2.155665777s" podCreationTimestamp="2025-09-29 21:25:36 +0000 UTC" firstStartedPulling="2025-09-29 21:25:37.078651078 +0000 UTC m=+8178.726440410" lastFinishedPulling="2025-09-29 21:25:37.61629815 +0000 UTC m=+8179.264087492" observedRunningTime="2025-09-29 21:25:38.153455307 +0000 UTC m=+8179.801244639" watchObservedRunningTime="2025-09-29 21:25:38.155665777 +0000 UTC m=+8179.803455099" Sep 29 21:25:42 crc kubenswrapper[4741]: I0929 21:25:42.087811 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:25:42 crc kubenswrapper[4741]: E0929 21:25:42.088687 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:25:53 crc kubenswrapper[4741]: I0929 21:25:53.085877 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:25:53 crc kubenswrapper[4741]: E0929 21:25:53.086743 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:26:08 crc kubenswrapper[4741]: I0929 21:26:08.086400 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:26:08 crc kubenswrapper[4741]: E0929 21:26:08.087360 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:26:22 crc kubenswrapper[4741]: I0929 21:26:22.085455 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:26:22 crc kubenswrapper[4741]: E0929 21:26:22.086265 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:26:33 crc kubenswrapper[4741]: I0929 21:26:33.087790 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:26:33 crc kubenswrapper[4741]: E0929 21:26:33.089517 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:26:47 crc kubenswrapper[4741]: I0929 21:26:47.087304 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:26:47 crc kubenswrapper[4741]: E0929 21:26:47.088052 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:27:01 crc kubenswrapper[4741]: I0929 21:27:01.086347 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:27:01 crc kubenswrapper[4741]: E0929 21:27:01.087330 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:27:13 crc kubenswrapper[4741]: I0929 21:27:13.085886 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:27:13 crc kubenswrapper[4741]: E0929 21:27:13.086674 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:27:25 crc kubenswrapper[4741]: I0929 21:27:25.086709 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:27:25 crc kubenswrapper[4741]: E0929 21:27:25.087352 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:27:37 crc kubenswrapper[4741]: I0929 21:27:37.086017 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:27:37 crc kubenswrapper[4741]: E0929 21:27:37.086705 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:27:49 crc kubenswrapper[4741]: I0929 21:27:49.087267 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:27:49 crc kubenswrapper[4741]: E0929 21:27:49.087951 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:28:00 crc kubenswrapper[4741]: I0929 21:28:00.086744 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:28:00 crc kubenswrapper[4741]: E0929 21:28:00.087715 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:28:13 crc kubenswrapper[4741]: I0929 21:28:13.090478 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:28:13 crc kubenswrapper[4741]: E0929 21:28:13.091330 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:28:26 crc kubenswrapper[4741]: I0929 21:28:26.085531 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:28:26 crc kubenswrapper[4741]: E0929 21:28:26.086265 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:28:40 crc kubenswrapper[4741]: I0929 21:28:40.086530 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:28:40 crc kubenswrapper[4741]: E0929 21:28:40.087407 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:28:53 crc kubenswrapper[4741]: I0929 21:28:53.087497 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:28:53 crc kubenswrapper[4741]: E0929 21:28:53.088288 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:29:04 crc kubenswrapper[4741]: I0929 21:29:04.086932 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:29:04 crc kubenswrapper[4741]: E0929 21:29:04.087728 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:29:18 crc kubenswrapper[4741]: I0929 21:29:18.086792 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:29:18 crc kubenswrapper[4741]: E0929 21:29:18.088047 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:29:31 crc kubenswrapper[4741]: I0929 21:29:31.108371 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:29:31 crc kubenswrapper[4741]: E0929 21:29:31.113899 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:29:44 crc kubenswrapper[4741]: I0929 21:29:44.086305 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:29:44 crc kubenswrapper[4741]: I0929 21:29:44.779522 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"93c64ec19ecaf263112fdbcd0c747acbe7b6f772ed4bd000171f227e7e688116"} Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.163814 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2"] Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.166202 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.170031 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.170711 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.178995 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2"] Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.371815 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e401cbbd-52df-46b0-96d9-a4d5144e2b02-config-volume\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.372185 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jc47\" (UniqueName: \"kubernetes.io/projected/e401cbbd-52df-46b0-96d9-a4d5144e2b02-kube-api-access-7jc47\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.372255 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e401cbbd-52df-46b0-96d9-a4d5144e2b02-secret-volume\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.474009 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e401cbbd-52df-46b0-96d9-a4d5144e2b02-config-volume\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.474119 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jc47\" (UniqueName: \"kubernetes.io/projected/e401cbbd-52df-46b0-96d9-a4d5144e2b02-kube-api-access-7jc47\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.474155 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e401cbbd-52df-46b0-96d9-a4d5144e2b02-secret-volume\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.474823 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e401cbbd-52df-46b0-96d9-a4d5144e2b02-config-volume\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.484197 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e401cbbd-52df-46b0-96d9-a4d5144e2b02-secret-volume\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.489278 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jc47\" (UniqueName: \"kubernetes.io/projected/e401cbbd-52df-46b0-96d9-a4d5144e2b02-kube-api-access-7jc47\") pod \"collect-profiles-29319690-b8cl2\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.501935 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:00 crc kubenswrapper[4741]: I0929 21:30:00.958038 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2"] Sep 29 21:30:01 crc kubenswrapper[4741]: I0929 21:30:01.971454 4741 generic.go:334] "Generic (PLEG): container finished" podID="e401cbbd-52df-46b0-96d9-a4d5144e2b02" containerID="b72163760202e65f21cc50d026c9239a76f7628d4cd489ced94bd365d4052c14" exitCode=0 Sep 29 21:30:01 crc kubenswrapper[4741]: I0929 21:30:01.971556 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" event={"ID":"e401cbbd-52df-46b0-96d9-a4d5144e2b02","Type":"ContainerDied","Data":"b72163760202e65f21cc50d026c9239a76f7628d4cd489ced94bd365d4052c14"} Sep 29 21:30:01 crc kubenswrapper[4741]: I0929 21:30:01.971841 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" event={"ID":"e401cbbd-52df-46b0-96d9-a4d5144e2b02","Type":"ContainerStarted","Data":"a889a1969eaf454bd8b5cbcd24d67f37ac700232998a18efefb57ad7cb4a3c5c"} Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.443686 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.537094 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e401cbbd-52df-46b0-96d9-a4d5144e2b02-config-volume\") pod \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.537433 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jc47\" (UniqueName: \"kubernetes.io/projected/e401cbbd-52df-46b0-96d9-a4d5144e2b02-kube-api-access-7jc47\") pod \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.537477 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e401cbbd-52df-46b0-96d9-a4d5144e2b02-secret-volume\") pod \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\" (UID: \"e401cbbd-52df-46b0-96d9-a4d5144e2b02\") " Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.537881 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e401cbbd-52df-46b0-96d9-a4d5144e2b02-config-volume" (OuterVolumeSpecName: "config-volume") pod "e401cbbd-52df-46b0-96d9-a4d5144e2b02" (UID: "e401cbbd-52df-46b0-96d9-a4d5144e2b02"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.538176 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e401cbbd-52df-46b0-96d9-a4d5144e2b02-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.542923 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e401cbbd-52df-46b0-96d9-a4d5144e2b02-kube-api-access-7jc47" (OuterVolumeSpecName: "kube-api-access-7jc47") pod "e401cbbd-52df-46b0-96d9-a4d5144e2b02" (UID: "e401cbbd-52df-46b0-96d9-a4d5144e2b02"). InnerVolumeSpecName "kube-api-access-7jc47". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.543069 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e401cbbd-52df-46b0-96d9-a4d5144e2b02-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e401cbbd-52df-46b0-96d9-a4d5144e2b02" (UID: "e401cbbd-52df-46b0-96d9-a4d5144e2b02"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.639594 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jc47\" (UniqueName: \"kubernetes.io/projected/e401cbbd-52df-46b0-96d9-a4d5144e2b02-kube-api-access-7jc47\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.639776 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e401cbbd-52df-46b0-96d9-a4d5144e2b02-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.997095 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.997107 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319690-b8cl2" event={"ID":"e401cbbd-52df-46b0-96d9-a4d5144e2b02","Type":"ContainerDied","Data":"a889a1969eaf454bd8b5cbcd24d67f37ac700232998a18efefb57ad7cb4a3c5c"} Sep 29 21:30:03 crc kubenswrapper[4741]: I0929 21:30:03.997147 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a889a1969eaf454bd8b5cbcd24d67f37ac700232998a18efefb57ad7cb4a3c5c" Sep 29 21:30:04 crc kubenswrapper[4741]: I0929 21:30:04.000553 4741 generic.go:334] "Generic (PLEG): container finished" podID="02f07af8-50b9-4252-90b9-a04657a5916d" containerID="6d4d2acd5bb12f72daf81add5ac54df6bfc49e19049a7c1dcbefafcd180bc8a9" exitCode=0 Sep 29 21:30:04 crc kubenswrapper[4741]: I0929 21:30:04.000648 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" event={"ID":"02f07af8-50b9-4252-90b9-a04657a5916d","Type":"ContainerDied","Data":"6d4d2acd5bb12f72daf81add5ac54df6bfc49e19049a7c1dcbefafcd180bc8a9"} Sep 29 21:30:04 crc kubenswrapper[4741]: I0929 21:30:04.524154 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj"] Sep 29 21:30:04 crc kubenswrapper[4741]: I0929 21:30:04.533571 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319645-2rvlj"] Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.098447 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b60a4868-d1e2-495a-9072-804ba359f428" path="/var/lib/kubelet/pods/b60a4868-d1e2-495a-9072-804ba359f428/volumes" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.425384 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.489974 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-2\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.490073 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-0\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.490115 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvgxq\" (UniqueName: \"kubernetes.io/projected/02f07af8-50b9-4252-90b9-a04657a5916d-kube-api-access-dvgxq\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.490222 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-inventory\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.490434 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ssh-key\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.490508 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-1\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.490553 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-telemetry-combined-ca-bundle\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.490591 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceph\") pod \"02f07af8-50b9-4252-90b9-a04657a5916d\" (UID: \"02f07af8-50b9-4252-90b9-a04657a5916d\") " Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.496998 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.497250 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceph" (OuterVolumeSpecName: "ceph") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.497305 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f07af8-50b9-4252-90b9-a04657a5916d-kube-api-access-dvgxq" (OuterVolumeSpecName: "kube-api-access-dvgxq") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "kube-api-access-dvgxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.521865 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.522293 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.524802 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-inventory" (OuterVolumeSpecName: "inventory") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.528281 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.530658 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "02f07af8-50b9-4252-90b9-a04657a5916d" (UID: "02f07af8-50b9-4252-90b9-a04657a5916d"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.591954 4741 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.591976 4741 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.591987 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvgxq\" (UniqueName: \"kubernetes.io/projected/02f07af8-50b9-4252-90b9-a04657a5916d-kube-api-access-dvgxq\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.591997 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.592005 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.592014 4741 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.592023 4741 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:05 crc kubenswrapper[4741]: I0929 21:30:05.592031 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02f07af8-50b9-4252-90b9-a04657a5916d-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.019092 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" event={"ID":"02f07af8-50b9-4252-90b9-a04657a5916d","Type":"ContainerDied","Data":"edd3eba569d1f58085684d47475aa103080ea0dd449690eed4afbd841c126118"} Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.019674 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edd3eba569d1f58085684d47475aa103080ea0dd449690eed4afbd841c126118" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.019207 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-28fg5" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.163606 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qdcpx"] Sep 29 21:30:06 crc kubenswrapper[4741]: E0929 21:30:06.164709 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e401cbbd-52df-46b0-96d9-a4d5144e2b02" containerName="collect-profiles" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.164842 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e401cbbd-52df-46b0-96d9-a4d5144e2b02" containerName="collect-profiles" Sep 29 21:30:06 crc kubenswrapper[4741]: E0929 21:30:06.165008 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f07af8-50b9-4252-90b9-a04657a5916d" containerName="telemetry-openstack-openstack-cell1" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.165103 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f07af8-50b9-4252-90b9-a04657a5916d" containerName="telemetry-openstack-openstack-cell1" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.165619 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e401cbbd-52df-46b0-96d9-a4d5144e2b02" containerName="collect-profiles" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.165757 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f07af8-50b9-4252-90b9-a04657a5916d" containerName="telemetry-openstack-openstack-cell1" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.167329 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.169603 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.171553 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.172026 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.172869 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.173349 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.195084 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qdcpx"] Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.203152 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.203222 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.203284 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.203316 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.203375 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4nfm\" (UniqueName: \"kubernetes.io/projected/19605586-361e-4c9a-9d60-685fd74ddb71-kube-api-access-t4nfm\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.203456 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.305749 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.306253 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.306683 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.306773 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.307012 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4nfm\" (UniqueName: \"kubernetes.io/projected/19605586-361e-4c9a-9d60-685fd74ddb71-kube-api-access-t4nfm\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.307207 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.312710 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.313085 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.317168 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.318186 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.320058 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.326580 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4nfm\" (UniqueName: \"kubernetes.io/projected/19605586-361e-4c9a-9d60-685fd74ddb71-kube-api-access-t4nfm\") pod \"neutron-sriov-openstack-openstack-cell1-qdcpx\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:06 crc kubenswrapper[4741]: I0929 21:30:06.494551 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:30:07 crc kubenswrapper[4741]: I0929 21:30:07.037090 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qdcpx"] Sep 29 21:30:07 crc kubenswrapper[4741]: W0929 21:30:07.043227 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19605586_361e_4c9a_9d60_685fd74ddb71.slice/crio-0637a5347c87194d4b3ab5f6367bd1222dcdf20473b4030d8434ad9df2ec5857 WatchSource:0}: Error finding container 0637a5347c87194d4b3ab5f6367bd1222dcdf20473b4030d8434ad9df2ec5857: Status 404 returned error can't find the container with id 0637a5347c87194d4b3ab5f6367bd1222dcdf20473b4030d8434ad9df2ec5857 Sep 29 21:30:07 crc kubenswrapper[4741]: I0929 21:30:07.046967 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:30:08 crc kubenswrapper[4741]: I0929 21:30:08.048604 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" event={"ID":"19605586-361e-4c9a-9d60-685fd74ddb71","Type":"ContainerStarted","Data":"d956c6677b4f62c82fc974a82ec74f940665eeff37063230ea042985c16750d2"} Sep 29 21:30:08 crc kubenswrapper[4741]: I0929 21:30:08.048932 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" event={"ID":"19605586-361e-4c9a-9d60-685fd74ddb71","Type":"ContainerStarted","Data":"0637a5347c87194d4b3ab5f6367bd1222dcdf20473b4030d8434ad9df2ec5857"} Sep 29 21:30:08 crc kubenswrapper[4741]: I0929 21:30:08.070256 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" podStartSLOduration=1.367856138 podStartE2EDuration="2.070233832s" podCreationTimestamp="2025-09-29 21:30:06 +0000 UTC" firstStartedPulling="2025-09-29 21:30:07.046770693 +0000 UTC m=+8448.694560025" lastFinishedPulling="2025-09-29 21:30:07.749148387 +0000 UTC m=+8449.396937719" observedRunningTime="2025-09-29 21:30:08.063959544 +0000 UTC m=+8449.711748876" watchObservedRunningTime="2025-09-29 21:30:08.070233832 +0000 UTC m=+8449.718023164" Sep 29 21:30:32 crc kubenswrapper[4741]: I0929 21:30:32.433481 4741 scope.go:117] "RemoveContainer" containerID="3ac79bbaa6afbb06d0adb6779f2fcaed485593c93f1cc1100dafa2950133b104" Sep 29 21:31:22 crc kubenswrapper[4741]: I0929 21:31:22.914457 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-27mrx"] Sep 29 21:31:22 crc kubenswrapper[4741]: I0929 21:31:22.922126 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:22 crc kubenswrapper[4741]: I0929 21:31:22.936863 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-27mrx"] Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.028226 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-catalog-content\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.028303 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-utilities\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.028347 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wphkc\" (UniqueName: \"kubernetes.io/projected/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-kube-api-access-wphkc\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.130664 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-catalog-content\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.130708 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-utilities\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.130728 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wphkc\" (UniqueName: \"kubernetes.io/projected/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-kube-api-access-wphkc\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.131212 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-catalog-content\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.131503 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-utilities\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.154116 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wphkc\" (UniqueName: \"kubernetes.io/projected/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-kube-api-access-wphkc\") pod \"certified-operators-27mrx\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.247889 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.804188 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-27mrx"] Sep 29 21:31:23 crc kubenswrapper[4741]: I0929 21:31:23.854699 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27mrx" event={"ID":"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a","Type":"ContainerStarted","Data":"8dac61b36bfba2ecef477f68eec16536cba12ac2b9be38673872bb386ec6a7d7"} Sep 29 21:31:24 crc kubenswrapper[4741]: I0929 21:31:24.865481 4741 generic.go:334] "Generic (PLEG): container finished" podID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerID="23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2" exitCode=0 Sep 29 21:31:24 crc kubenswrapper[4741]: I0929 21:31:24.866463 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27mrx" event={"ID":"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a","Type":"ContainerDied","Data":"23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2"} Sep 29 21:31:26 crc kubenswrapper[4741]: I0929 21:31:26.895066 4741 generic.go:334] "Generic (PLEG): container finished" podID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerID="9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603" exitCode=0 Sep 29 21:31:26 crc kubenswrapper[4741]: I0929 21:31:26.895154 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27mrx" event={"ID":"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a","Type":"ContainerDied","Data":"9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603"} Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.696493 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b5hgf"] Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.699491 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.706087 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5hgf"] Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.829603 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-catalog-content\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.829657 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-utilities\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.829698 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b42cn\" (UniqueName: \"kubernetes.io/projected/61fe09a1-ef05-425f-a381-864ae8ca8eee-kube-api-access-b42cn\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.907725 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27mrx" event={"ID":"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a","Type":"ContainerStarted","Data":"8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c"} Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.931416 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-catalog-content\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.931522 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-utilities\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.931628 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b42cn\" (UniqueName: \"kubernetes.io/projected/61fe09a1-ef05-425f-a381-864ae8ca8eee-kube-api-access-b42cn\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.932481 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-catalog-content\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.932481 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-utilities\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:27 crc kubenswrapper[4741]: I0929 21:31:27.961535 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b42cn\" (UniqueName: \"kubernetes.io/projected/61fe09a1-ef05-425f-a381-864ae8ca8eee-kube-api-access-b42cn\") pod \"redhat-marketplace-b5hgf\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.073119 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.085079 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-27mrx" podStartSLOduration=3.418191771 podStartE2EDuration="6.085062067s" podCreationTimestamp="2025-09-29 21:31:22 +0000 UTC" firstStartedPulling="2025-09-29 21:31:24.868514089 +0000 UTC m=+8526.516303421" lastFinishedPulling="2025-09-29 21:31:27.535384385 +0000 UTC m=+8529.183173717" observedRunningTime="2025-09-29 21:31:27.937750505 +0000 UTC m=+8529.585539837" watchObservedRunningTime="2025-09-29 21:31:28.085062067 +0000 UTC m=+8529.732851399" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.088114 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-np2bt"] Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.090322 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.103719 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-np2bt"] Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.245646 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94jsr\" (UniqueName: \"kubernetes.io/projected/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-kube-api-access-94jsr\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.245986 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-catalog-content\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.246276 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-utilities\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.347641 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94jsr\" (UniqueName: \"kubernetes.io/projected/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-kube-api-access-94jsr\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.347688 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-catalog-content\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.347826 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-utilities\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.348268 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-utilities\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.348855 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-catalog-content\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.372914 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94jsr\" (UniqueName: \"kubernetes.io/projected/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-kube-api-access-94jsr\") pod \"redhat-operators-np2bt\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.563978 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.585991 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5hgf"] Sep 29 21:31:28 crc kubenswrapper[4741]: W0929 21:31:28.602179 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61fe09a1_ef05_425f_a381_864ae8ca8eee.slice/crio-6770c41f3f8231265bdaf11bcb9a0479b3671aa312ea16431b6c2d62f211d35c WatchSource:0}: Error finding container 6770c41f3f8231265bdaf11bcb9a0479b3671aa312ea16431b6c2d62f211d35c: Status 404 returned error can't find the container with id 6770c41f3f8231265bdaf11bcb9a0479b3671aa312ea16431b6c2d62f211d35c Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.922659 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5hgf" event={"ID":"61fe09a1-ef05-425f-a381-864ae8ca8eee","Type":"ContainerStarted","Data":"3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30"} Sep 29 21:31:28 crc kubenswrapper[4741]: I0929 21:31:28.922876 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5hgf" event={"ID":"61fe09a1-ef05-425f-a381-864ae8ca8eee","Type":"ContainerStarted","Data":"6770c41f3f8231265bdaf11bcb9a0479b3671aa312ea16431b6c2d62f211d35c"} Sep 29 21:31:29 crc kubenswrapper[4741]: I0929 21:31:29.142140 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-np2bt"] Sep 29 21:31:29 crc kubenswrapper[4741]: I0929 21:31:29.931966 4741 generic.go:334] "Generic (PLEG): container finished" podID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerID="0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60" exitCode=0 Sep 29 21:31:29 crc kubenswrapper[4741]: I0929 21:31:29.932063 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np2bt" event={"ID":"1bdf4ef2-3408-4342-b411-5a5ed9d5b781","Type":"ContainerDied","Data":"0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60"} Sep 29 21:31:29 crc kubenswrapper[4741]: I0929 21:31:29.932308 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np2bt" event={"ID":"1bdf4ef2-3408-4342-b411-5a5ed9d5b781","Type":"ContainerStarted","Data":"b313d252c81bf83ceee558bfff10becd848f973b959d3ab56ec7eb2afa9996ab"} Sep 29 21:31:29 crc kubenswrapper[4741]: I0929 21:31:29.935656 4741 generic.go:334] "Generic (PLEG): container finished" podID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerID="3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30" exitCode=0 Sep 29 21:31:29 crc kubenswrapper[4741]: I0929 21:31:29.935890 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5hgf" event={"ID":"61fe09a1-ef05-425f-a381-864ae8ca8eee","Type":"ContainerDied","Data":"3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30"} Sep 29 21:31:30 crc kubenswrapper[4741]: I0929 21:31:30.947720 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np2bt" event={"ID":"1bdf4ef2-3408-4342-b411-5a5ed9d5b781","Type":"ContainerStarted","Data":"dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18"} Sep 29 21:31:31 crc kubenswrapper[4741]: I0929 21:31:31.958245 4741 generic.go:334] "Generic (PLEG): container finished" podID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerID="1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25" exitCode=0 Sep 29 21:31:31 crc kubenswrapper[4741]: I0929 21:31:31.958370 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5hgf" event={"ID":"61fe09a1-ef05-425f-a381-864ae8ca8eee","Type":"ContainerDied","Data":"1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25"} Sep 29 21:31:32 crc kubenswrapper[4741]: I0929 21:31:32.992060 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5hgf" event={"ID":"61fe09a1-ef05-425f-a381-864ae8ca8eee","Type":"ContainerStarted","Data":"cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f"} Sep 29 21:31:33 crc kubenswrapper[4741]: I0929 21:31:33.020196 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b5hgf" podStartSLOduration=3.404251351 podStartE2EDuration="6.020177817s" podCreationTimestamp="2025-09-29 21:31:27 +0000 UTC" firstStartedPulling="2025-09-29 21:31:29.945743546 +0000 UTC m=+8531.593532878" lastFinishedPulling="2025-09-29 21:31:32.561670012 +0000 UTC m=+8534.209459344" observedRunningTime="2025-09-29 21:31:33.014452697 +0000 UTC m=+8534.662242049" watchObservedRunningTime="2025-09-29 21:31:33.020177817 +0000 UTC m=+8534.667967149" Sep 29 21:31:33 crc kubenswrapper[4741]: I0929 21:31:33.248748 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:33 crc kubenswrapper[4741]: I0929 21:31:33.249104 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:33 crc kubenswrapper[4741]: I0929 21:31:33.311602 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:34 crc kubenswrapper[4741]: I0929 21:31:34.007323 4741 generic.go:334] "Generic (PLEG): container finished" podID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerID="dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18" exitCode=0 Sep 29 21:31:34 crc kubenswrapper[4741]: I0929 21:31:34.008217 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np2bt" event={"ID":"1bdf4ef2-3408-4342-b411-5a5ed9d5b781","Type":"ContainerDied","Data":"dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18"} Sep 29 21:31:34 crc kubenswrapper[4741]: I0929 21:31:34.062166 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:35 crc kubenswrapper[4741]: I0929 21:31:35.019592 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np2bt" event={"ID":"1bdf4ef2-3408-4342-b411-5a5ed9d5b781","Type":"ContainerStarted","Data":"f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89"} Sep 29 21:31:35 crc kubenswrapper[4741]: I0929 21:31:35.037516 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-np2bt" podStartSLOduration=2.438211189 podStartE2EDuration="7.037493181s" podCreationTimestamp="2025-09-29 21:31:28 +0000 UTC" firstStartedPulling="2025-09-29 21:31:29.933619075 +0000 UTC m=+8531.581408407" lastFinishedPulling="2025-09-29 21:31:34.532901067 +0000 UTC m=+8536.180690399" observedRunningTime="2025-09-29 21:31:35.035171958 +0000 UTC m=+8536.682961300" watchObservedRunningTime="2025-09-29 21:31:35.037493181 +0000 UTC m=+8536.685282513" Sep 29 21:31:36 crc kubenswrapper[4741]: I0929 21:31:36.880285 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-27mrx"] Sep 29 21:31:36 crc kubenswrapper[4741]: I0929 21:31:36.880782 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-27mrx" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="registry-server" containerID="cri-o://8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c" gracePeriod=2 Sep 29 21:31:37 crc kubenswrapper[4741]: I0929 21:31:37.864525 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:37 crc kubenswrapper[4741]: I0929 21:31:37.993326 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-utilities\") pod \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " Sep 29 21:31:37 crc kubenswrapper[4741]: I0929 21:31:37.993372 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-catalog-content\") pod \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " Sep 29 21:31:37 crc kubenswrapper[4741]: I0929 21:31:37.993418 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wphkc\" (UniqueName: \"kubernetes.io/projected/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-kube-api-access-wphkc\") pod \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\" (UID: \"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a\") " Sep 29 21:31:37 crc kubenswrapper[4741]: I0929 21:31:37.994356 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-utilities" (OuterVolumeSpecName: "utilities") pod "5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" (UID: "5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.002068 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-kube-api-access-wphkc" (OuterVolumeSpecName: "kube-api-access-wphkc") pod "5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" (UID: "5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a"). InnerVolumeSpecName "kube-api-access-wphkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.036564 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" (UID: "5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.049908 4741 generic.go:334] "Generic (PLEG): container finished" podID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerID="8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c" exitCode=0 Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.049962 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27mrx" event={"ID":"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a","Type":"ContainerDied","Data":"8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c"} Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.049993 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27mrx" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.050012 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27mrx" event={"ID":"5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a","Type":"ContainerDied","Data":"8dac61b36bfba2ecef477f68eec16536cba12ac2b9be38673872bb386ec6a7d7"} Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.050031 4741 scope.go:117] "RemoveContainer" containerID="8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.073350 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.073641 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.081362 4741 scope.go:117] "RemoveContainer" containerID="9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.081383 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-27mrx"] Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.089565 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-27mrx"] Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.096578 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.096619 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.096682 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wphkc\" (UniqueName: \"kubernetes.io/projected/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a-kube-api-access-wphkc\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.106685 4741 scope.go:117] "RemoveContainer" containerID="23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.136996 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.159366 4741 scope.go:117] "RemoveContainer" containerID="8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c" Sep 29 21:31:38 crc kubenswrapper[4741]: E0929 21:31:38.159789 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c\": container with ID starting with 8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c not found: ID does not exist" containerID="8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.159850 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c"} err="failed to get container status \"8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c\": rpc error: code = NotFound desc = could not find container \"8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c\": container with ID starting with 8b7544df7a77a47bd77f9ad1a35b82d6c0251403d332c267ed95b8658f4bdd0c not found: ID does not exist" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.159878 4741 scope.go:117] "RemoveContainer" containerID="9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603" Sep 29 21:31:38 crc kubenswrapper[4741]: E0929 21:31:38.160179 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603\": container with ID starting with 9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603 not found: ID does not exist" containerID="9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.160203 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603"} err="failed to get container status \"9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603\": rpc error: code = NotFound desc = could not find container \"9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603\": container with ID starting with 9ac1b7c7d80dd228e5519a8ed09b88e48207a1081613dedcc7024a952ffd2603 not found: ID does not exist" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.160219 4741 scope.go:117] "RemoveContainer" containerID="23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2" Sep 29 21:31:38 crc kubenswrapper[4741]: E0929 21:31:38.160499 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2\": container with ID starting with 23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2 not found: ID does not exist" containerID="23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.160522 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2"} err="failed to get container status \"23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2\": rpc error: code = NotFound desc = could not find container \"23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2\": container with ID starting with 23347cee6a12da99fa98ccffe90e9eb99d39b858f70270287f32ab9e844fb9a2 not found: ID does not exist" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.564440 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:38 crc kubenswrapper[4741]: I0929 21:31:38.564694 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:39 crc kubenswrapper[4741]: I0929 21:31:39.097490 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" path="/var/lib/kubelet/pods/5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a/volumes" Sep 29 21:31:39 crc kubenswrapper[4741]: I0929 21:31:39.114662 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:39 crc kubenswrapper[4741]: I0929 21:31:39.609812 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-np2bt" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="registry-server" probeResult="failure" output=< Sep 29 21:31:39 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 21:31:39 crc kubenswrapper[4741]: > Sep 29 21:31:40 crc kubenswrapper[4741]: I0929 21:31:40.485250 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5hgf"] Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.132886 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b5hgf" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="registry-server" containerID="cri-o://cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f" gracePeriod=2 Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.690137 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.802043 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-utilities\") pod \"61fe09a1-ef05-425f-a381-864ae8ca8eee\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.802466 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b42cn\" (UniqueName: \"kubernetes.io/projected/61fe09a1-ef05-425f-a381-864ae8ca8eee-kube-api-access-b42cn\") pod \"61fe09a1-ef05-425f-a381-864ae8ca8eee\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.802500 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-catalog-content\") pod \"61fe09a1-ef05-425f-a381-864ae8ca8eee\" (UID: \"61fe09a1-ef05-425f-a381-864ae8ca8eee\") " Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.812966 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61fe09a1-ef05-425f-a381-864ae8ca8eee-kube-api-access-b42cn" (OuterVolumeSpecName: "kube-api-access-b42cn") pod "61fe09a1-ef05-425f-a381-864ae8ca8eee" (UID: "61fe09a1-ef05-425f-a381-864ae8ca8eee"). InnerVolumeSpecName "kube-api-access-b42cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.815047 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-utilities" (OuterVolumeSpecName: "utilities") pod "61fe09a1-ef05-425f-a381-864ae8ca8eee" (UID: "61fe09a1-ef05-425f-a381-864ae8ca8eee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.815652 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61fe09a1-ef05-425f-a381-864ae8ca8eee" (UID: "61fe09a1-ef05-425f-a381-864ae8ca8eee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.904763 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b42cn\" (UniqueName: \"kubernetes.io/projected/61fe09a1-ef05-425f-a381-864ae8ca8eee-kube-api-access-b42cn\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.904796 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:42 crc kubenswrapper[4741]: I0929 21:31:42.904805 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fe09a1-ef05-425f-a381-864ae8ca8eee-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.146195 4741 generic.go:334] "Generic (PLEG): container finished" podID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerID="cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f" exitCode=0 Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.146233 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5hgf" event={"ID":"61fe09a1-ef05-425f-a381-864ae8ca8eee","Type":"ContainerDied","Data":"cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f"} Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.146259 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5hgf" event={"ID":"61fe09a1-ef05-425f-a381-864ae8ca8eee","Type":"ContainerDied","Data":"6770c41f3f8231265bdaf11bcb9a0479b3671aa312ea16431b6c2d62f211d35c"} Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.146274 4741 scope.go:117] "RemoveContainer" containerID="cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.146433 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5hgf" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.178067 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5hgf"] Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.182009 4741 scope.go:117] "RemoveContainer" containerID="1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.191145 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5hgf"] Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.201850 4741 scope.go:117] "RemoveContainer" containerID="3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.295404 4741 scope.go:117] "RemoveContainer" containerID="cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f" Sep 29 21:31:43 crc kubenswrapper[4741]: E0929 21:31:43.296156 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f\": container with ID starting with cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f not found: ID does not exist" containerID="cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.296202 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f"} err="failed to get container status \"cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f\": rpc error: code = NotFound desc = could not find container \"cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f\": container with ID starting with cfeeffd75ca175f9c9a87d1e920ec374bcc32280852e1fa1747c74e3bb767c4f not found: ID does not exist" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.296302 4741 scope.go:117] "RemoveContainer" containerID="1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25" Sep 29 21:31:43 crc kubenswrapper[4741]: E0929 21:31:43.297100 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25\": container with ID starting with 1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25 not found: ID does not exist" containerID="1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.297150 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25"} err="failed to get container status \"1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25\": rpc error: code = NotFound desc = could not find container \"1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25\": container with ID starting with 1b8ff6c005e856146ed34f91048a1e377e5f2dcc56e9b62ef5ffbc2ea1e57a25 not found: ID does not exist" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.297179 4741 scope.go:117] "RemoveContainer" containerID="3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30" Sep 29 21:31:43 crc kubenswrapper[4741]: E0929 21:31:43.297650 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30\": container with ID starting with 3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30 not found: ID does not exist" containerID="3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30" Sep 29 21:31:43 crc kubenswrapper[4741]: I0929 21:31:43.297694 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30"} err="failed to get container status \"3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30\": rpc error: code = NotFound desc = could not find container \"3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30\": container with ID starting with 3c501a3d822b5d49805f0ebae4a8056452847bc5c55466ba1b242e1f20b82f30 not found: ID does not exist" Sep 29 21:31:45 crc kubenswrapper[4741]: I0929 21:31:45.098279 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" path="/var/lib/kubelet/pods/61fe09a1-ef05-425f-a381-864ae8ca8eee/volumes" Sep 29 21:31:48 crc kubenswrapper[4741]: I0929 21:31:48.616046 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:48 crc kubenswrapper[4741]: I0929 21:31:48.675011 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:48 crc kubenswrapper[4741]: I0929 21:31:48.853843 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-np2bt"] Sep 29 21:31:50 crc kubenswrapper[4741]: I0929 21:31:50.228567 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-np2bt" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="registry-server" containerID="cri-o://f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89" gracePeriod=2 Sep 29 21:31:50 crc kubenswrapper[4741]: I0929 21:31:50.797347 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:50 crc kubenswrapper[4741]: I0929 21:31:50.995638 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-catalog-content\") pod \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " Sep 29 21:31:50 crc kubenswrapper[4741]: I0929 21:31:50.995726 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94jsr\" (UniqueName: \"kubernetes.io/projected/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-kube-api-access-94jsr\") pod \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " Sep 29 21:31:50 crc kubenswrapper[4741]: I0929 21:31:50.995817 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-utilities\") pod \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\" (UID: \"1bdf4ef2-3408-4342-b411-5a5ed9d5b781\") " Sep 29 21:31:50 crc kubenswrapper[4741]: I0929 21:31:50.997129 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-utilities" (OuterVolumeSpecName: "utilities") pod "1bdf4ef2-3408-4342-b411-5a5ed9d5b781" (UID: "1bdf4ef2-3408-4342-b411-5a5ed9d5b781"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.007732 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-kube-api-access-94jsr" (OuterVolumeSpecName: "kube-api-access-94jsr") pod "1bdf4ef2-3408-4342-b411-5a5ed9d5b781" (UID: "1bdf4ef2-3408-4342-b411-5a5ed9d5b781"). InnerVolumeSpecName "kube-api-access-94jsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.083363 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bdf4ef2-3408-4342-b411-5a5ed9d5b781" (UID: "1bdf4ef2-3408-4342-b411-5a5ed9d5b781"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.099235 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.099280 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94jsr\" (UniqueName: \"kubernetes.io/projected/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-kube-api-access-94jsr\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.099296 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdf4ef2-3408-4342-b411-5a5ed9d5b781-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.243148 4741 generic.go:334] "Generic (PLEG): container finished" podID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerID="f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89" exitCode=0 Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.243196 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np2bt" event={"ID":"1bdf4ef2-3408-4342-b411-5a5ed9d5b781","Type":"ContainerDied","Data":"f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89"} Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.243229 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np2bt" event={"ID":"1bdf4ef2-3408-4342-b411-5a5ed9d5b781","Type":"ContainerDied","Data":"b313d252c81bf83ceee558bfff10becd848f973b959d3ab56ec7eb2afa9996ab"} Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.243252 4741 scope.go:117] "RemoveContainer" containerID="f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.243253 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np2bt" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.267818 4741 scope.go:117] "RemoveContainer" containerID="dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.270823 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-np2bt"] Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.278982 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-np2bt"] Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.293466 4741 scope.go:117] "RemoveContainer" containerID="0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.351624 4741 scope.go:117] "RemoveContainer" containerID="f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89" Sep 29 21:31:51 crc kubenswrapper[4741]: E0929 21:31:51.352154 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89\": container with ID starting with f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89 not found: ID does not exist" containerID="f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.352205 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89"} err="failed to get container status \"f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89\": rpc error: code = NotFound desc = could not find container \"f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89\": container with ID starting with f72d2833c0c6b2eb84eb421f0bdc1773657a69dabca57b0cacfe8377502a9b89 not found: ID does not exist" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.352239 4741 scope.go:117] "RemoveContainer" containerID="dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18" Sep 29 21:31:51 crc kubenswrapper[4741]: E0929 21:31:51.353051 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18\": container with ID starting with dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18 not found: ID does not exist" containerID="dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.353098 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18"} err="failed to get container status \"dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18\": rpc error: code = NotFound desc = could not find container \"dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18\": container with ID starting with dffb6a8213873f8c275c2a21808fddcfbcec10ead9f38fa251252000e4cf8e18 not found: ID does not exist" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.353129 4741 scope.go:117] "RemoveContainer" containerID="0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60" Sep 29 21:31:51 crc kubenswrapper[4741]: E0929 21:31:51.353420 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60\": container with ID starting with 0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60 not found: ID does not exist" containerID="0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60" Sep 29 21:31:51 crc kubenswrapper[4741]: I0929 21:31:51.353445 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60"} err="failed to get container status \"0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60\": rpc error: code = NotFound desc = could not find container \"0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60\": container with ID starting with 0dccc64fefac1e0f2da5362820e074c24e4b30c9d456ec1251c74e564899fd60 not found: ID does not exist" Sep 29 21:31:53 crc kubenswrapper[4741]: I0929 21:31:53.104503 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" path="/var/lib/kubelet/pods/1bdf4ef2-3408-4342-b411-5a5ed9d5b781/volumes" Sep 29 21:32:01 crc kubenswrapper[4741]: I0929 21:32:01.739013 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:32:01 crc kubenswrapper[4741]: I0929 21:32:01.739558 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:32:31 crc kubenswrapper[4741]: I0929 21:32:31.738594 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:32:31 crc kubenswrapper[4741]: I0929 21:32:31.739119 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.739088 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.739655 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.739693 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.740248 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93c64ec19ecaf263112fdbcd0c747acbe7b6f772ed4bd000171f227e7e688116"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.740291 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://93c64ec19ecaf263112fdbcd0c747acbe7b6f772ed4bd000171f227e7e688116" gracePeriod=600 Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.935021 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="93c64ec19ecaf263112fdbcd0c747acbe7b6f772ed4bd000171f227e7e688116" exitCode=0 Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.935221 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"93c64ec19ecaf263112fdbcd0c747acbe7b6f772ed4bd000171f227e7e688116"} Sep 29 21:33:01 crc kubenswrapper[4741]: I0929 21:33:01.935404 4741 scope.go:117] "RemoveContainer" containerID="f334c85ce09c1916f65b216ad35213a5015fa75b9c06d03ebda563f16fbeb43c" Sep 29 21:33:02 crc kubenswrapper[4741]: I0929 21:33:02.949442 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2"} Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.071705 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-27lxc"] Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072584 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="extract-utilities" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072597 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="extract-utilities" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072612 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="extract-content" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072618 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="extract-content" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072629 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="extract-utilities" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072635 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="extract-utilities" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072656 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="extract-content" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072662 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="extract-content" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072675 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="extract-content" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072681 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="extract-content" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072690 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="extract-utilities" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072695 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="extract-utilities" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072704 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072709 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072721 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072727 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: E0929 21:33:52.072740 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072746 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072956 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f5b8b7d-ea2d-476e-ab0a-7c39cdbcb02a" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072972 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bdf4ef2-3408-4342-b411-5a5ed9d5b781" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.072989 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fe09a1-ef05-425f-a381-864ae8ca8eee" containerName="registry-server" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.074718 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.082913 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-27lxc"] Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.173805 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-utilities\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.173860 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-catalog-content\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.174002 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk5lm\" (UniqueName: \"kubernetes.io/projected/23258818-8ba5-40a1-95f7-fe74390b1293-kube-api-access-wk5lm\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.276482 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-utilities\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.276539 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-catalog-content\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.276652 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk5lm\" (UniqueName: \"kubernetes.io/projected/23258818-8ba5-40a1-95f7-fe74390b1293-kube-api-access-wk5lm\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.276992 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-utilities\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.277079 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-catalog-content\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.312745 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk5lm\" (UniqueName: \"kubernetes.io/projected/23258818-8ba5-40a1-95f7-fe74390b1293-kube-api-access-wk5lm\") pod \"community-operators-27lxc\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.395588 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:33:52 crc kubenswrapper[4741]: I0929 21:33:52.892511 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-27lxc"] Sep 29 21:33:53 crc kubenswrapper[4741]: I0929 21:33:53.455098 4741 generic.go:334] "Generic (PLEG): container finished" podID="23258818-8ba5-40a1-95f7-fe74390b1293" containerID="e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233" exitCode=0 Sep 29 21:33:53 crc kubenswrapper[4741]: I0929 21:33:53.455149 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-27lxc" event={"ID":"23258818-8ba5-40a1-95f7-fe74390b1293","Type":"ContainerDied","Data":"e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233"} Sep 29 21:33:53 crc kubenswrapper[4741]: I0929 21:33:53.455467 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-27lxc" event={"ID":"23258818-8ba5-40a1-95f7-fe74390b1293","Type":"ContainerStarted","Data":"bb1dec1de6962abf1040fea3ef52d36d12a38cc1f0427197b930b18b0fc3caca"} Sep 29 21:33:55 crc kubenswrapper[4741]: I0929 21:33:55.477404 4741 generic.go:334] "Generic (PLEG): container finished" podID="23258818-8ba5-40a1-95f7-fe74390b1293" containerID="920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00" exitCode=0 Sep 29 21:33:55 crc kubenswrapper[4741]: I0929 21:33:55.477494 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-27lxc" event={"ID":"23258818-8ba5-40a1-95f7-fe74390b1293","Type":"ContainerDied","Data":"920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00"} Sep 29 21:33:56 crc kubenswrapper[4741]: I0929 21:33:56.489124 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-27lxc" event={"ID":"23258818-8ba5-40a1-95f7-fe74390b1293","Type":"ContainerStarted","Data":"e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc"} Sep 29 21:34:02 crc kubenswrapper[4741]: I0929 21:34:02.396098 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:34:02 crc kubenswrapper[4741]: I0929 21:34:02.396685 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:34:02 crc kubenswrapper[4741]: I0929 21:34:02.450330 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:34:02 crc kubenswrapper[4741]: I0929 21:34:02.475447 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-27lxc" podStartSLOduration=8.062180725 podStartE2EDuration="10.475425538s" podCreationTimestamp="2025-09-29 21:33:52 +0000 UTC" firstStartedPulling="2025-09-29 21:33:53.457439773 +0000 UTC m=+8675.105229125" lastFinishedPulling="2025-09-29 21:33:55.870684605 +0000 UTC m=+8677.518473938" observedRunningTime="2025-09-29 21:33:56.50694771 +0000 UTC m=+8678.154737042" watchObservedRunningTime="2025-09-29 21:34:02.475425538 +0000 UTC m=+8684.123214870" Sep 29 21:34:02 crc kubenswrapper[4741]: I0929 21:34:02.611214 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:34:02 crc kubenswrapper[4741]: I0929 21:34:02.688784 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-27lxc"] Sep 29 21:34:04 crc kubenswrapper[4741]: I0929 21:34:04.585631 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-27lxc" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="registry-server" containerID="cri-o://e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc" gracePeriod=2 Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.104377 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.177519 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-utilities\") pod \"23258818-8ba5-40a1-95f7-fe74390b1293\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.177724 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-catalog-content\") pod \"23258818-8ba5-40a1-95f7-fe74390b1293\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.177881 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk5lm\" (UniqueName: \"kubernetes.io/projected/23258818-8ba5-40a1-95f7-fe74390b1293-kube-api-access-wk5lm\") pod \"23258818-8ba5-40a1-95f7-fe74390b1293\" (UID: \"23258818-8ba5-40a1-95f7-fe74390b1293\") " Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.179062 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-utilities" (OuterVolumeSpecName: "utilities") pod "23258818-8ba5-40a1-95f7-fe74390b1293" (UID: "23258818-8ba5-40a1-95f7-fe74390b1293"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.198506 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23258818-8ba5-40a1-95f7-fe74390b1293-kube-api-access-wk5lm" (OuterVolumeSpecName: "kube-api-access-wk5lm") pod "23258818-8ba5-40a1-95f7-fe74390b1293" (UID: "23258818-8ba5-40a1-95f7-fe74390b1293"). InnerVolumeSpecName "kube-api-access-wk5lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.237473 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23258818-8ba5-40a1-95f7-fe74390b1293" (UID: "23258818-8ba5-40a1-95f7-fe74390b1293"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.280731 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk5lm\" (UniqueName: \"kubernetes.io/projected/23258818-8ba5-40a1-95f7-fe74390b1293-kube-api-access-wk5lm\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.280760 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.280773 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23258818-8ba5-40a1-95f7-fe74390b1293-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.598623 4741 generic.go:334] "Generic (PLEG): container finished" podID="23258818-8ba5-40a1-95f7-fe74390b1293" containerID="e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc" exitCode=0 Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.598659 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-27lxc" event={"ID":"23258818-8ba5-40a1-95f7-fe74390b1293","Type":"ContainerDied","Data":"e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc"} Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.598687 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-27lxc" event={"ID":"23258818-8ba5-40a1-95f7-fe74390b1293","Type":"ContainerDied","Data":"bb1dec1de6962abf1040fea3ef52d36d12a38cc1f0427197b930b18b0fc3caca"} Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.598701 4741 scope.go:117] "RemoveContainer" containerID="e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.598812 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-27lxc" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.638567 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-27lxc"] Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.647051 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-27lxc"] Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.648809 4741 scope.go:117] "RemoveContainer" containerID="920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.675146 4741 scope.go:117] "RemoveContainer" containerID="e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.724072 4741 scope.go:117] "RemoveContainer" containerID="e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc" Sep 29 21:34:05 crc kubenswrapper[4741]: E0929 21:34:05.724523 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc\": container with ID starting with e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc not found: ID does not exist" containerID="e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.724671 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc"} err="failed to get container status \"e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc\": rpc error: code = NotFound desc = could not find container \"e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc\": container with ID starting with e96816a730d65069335934f7651aa1bf43b8a987065adeacd3a8c8b38be875fc not found: ID does not exist" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.724774 4741 scope.go:117] "RemoveContainer" containerID="920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00" Sep 29 21:34:05 crc kubenswrapper[4741]: E0929 21:34:05.725404 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00\": container with ID starting with 920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00 not found: ID does not exist" containerID="920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.725514 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00"} err="failed to get container status \"920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00\": rpc error: code = NotFound desc = could not find container \"920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00\": container with ID starting with 920696b7a31a53ccfe641acf0243bbba8bfcf92591f64cef90985b2370d37f00 not found: ID does not exist" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.725597 4741 scope.go:117] "RemoveContainer" containerID="e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233" Sep 29 21:34:05 crc kubenswrapper[4741]: E0929 21:34:05.725914 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233\": container with ID starting with e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233 not found: ID does not exist" containerID="e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233" Sep 29 21:34:05 crc kubenswrapper[4741]: I0929 21:34:05.725950 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233"} err="failed to get container status \"e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233\": rpc error: code = NotFound desc = could not find container \"e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233\": container with ID starting with e7ba2310e628f01da87862023ab09746ab2a1342d829b9e4e2a9b92e44cf5233 not found: ID does not exist" Sep 29 21:34:07 crc kubenswrapper[4741]: I0929 21:34:07.099520 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" path="/var/lib/kubelet/pods/23258818-8ba5-40a1-95f7-fe74390b1293/volumes" Sep 29 21:34:17 crc kubenswrapper[4741]: I0929 21:34:17.730296 4741 generic.go:334] "Generic (PLEG): container finished" podID="19605586-361e-4c9a-9d60-685fd74ddb71" containerID="d956c6677b4f62c82fc974a82ec74f940665eeff37063230ea042985c16750d2" exitCode=0 Sep 29 21:34:17 crc kubenswrapper[4741]: I0929 21:34:17.730380 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" event={"ID":"19605586-361e-4c9a-9d60-685fd74ddb71","Type":"ContainerDied","Data":"d956c6677b4f62c82fc974a82ec74f940665eeff37063230ea042985c16750d2"} Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.404195 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.496604 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-combined-ca-bundle\") pod \"19605586-361e-4c9a-9d60-685fd74ddb71\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.496661 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ssh-key\") pod \"19605586-361e-4c9a-9d60-685fd74ddb71\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.496745 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-inventory\") pod \"19605586-361e-4c9a-9d60-685fd74ddb71\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.496842 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4nfm\" (UniqueName: \"kubernetes.io/projected/19605586-361e-4c9a-9d60-685fd74ddb71-kube-api-access-t4nfm\") pod \"19605586-361e-4c9a-9d60-685fd74ddb71\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.496903 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ceph\") pod \"19605586-361e-4c9a-9d60-685fd74ddb71\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.496961 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-agent-neutron-config-0\") pod \"19605586-361e-4c9a-9d60-685fd74ddb71\" (UID: \"19605586-361e-4c9a-9d60-685fd74ddb71\") " Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.503511 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "19605586-361e-4c9a-9d60-685fd74ddb71" (UID: "19605586-361e-4c9a-9d60-685fd74ddb71"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.503573 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ceph" (OuterVolumeSpecName: "ceph") pod "19605586-361e-4c9a-9d60-685fd74ddb71" (UID: "19605586-361e-4c9a-9d60-685fd74ddb71"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.503570 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19605586-361e-4c9a-9d60-685fd74ddb71-kube-api-access-t4nfm" (OuterVolumeSpecName: "kube-api-access-t4nfm") pod "19605586-361e-4c9a-9d60-685fd74ddb71" (UID: "19605586-361e-4c9a-9d60-685fd74ddb71"). InnerVolumeSpecName "kube-api-access-t4nfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.528829 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19605586-361e-4c9a-9d60-685fd74ddb71" (UID: "19605586-361e-4c9a-9d60-685fd74ddb71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.529116 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "19605586-361e-4c9a-9d60-685fd74ddb71" (UID: "19605586-361e-4c9a-9d60-685fd74ddb71"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.535543 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-inventory" (OuterVolumeSpecName: "inventory") pod "19605586-361e-4c9a-9d60-685fd74ddb71" (UID: "19605586-361e-4c9a-9d60-685fd74ddb71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.599430 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.599463 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4nfm\" (UniqueName: \"kubernetes.io/projected/19605586-361e-4c9a-9d60-685fd74ddb71-kube-api-access-t4nfm\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.599480 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.599489 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.599500 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.599508 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19605586-361e-4c9a-9d60-685fd74ddb71-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.760218 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" event={"ID":"19605586-361e-4c9a-9d60-685fd74ddb71","Type":"ContainerDied","Data":"0637a5347c87194d4b3ab5f6367bd1222dcdf20473b4030d8434ad9df2ec5857"} Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.760258 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0637a5347c87194d4b3ab5f6367bd1222dcdf20473b4030d8434ad9df2ec5857" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.760347 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qdcpx" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.908875 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg"] Sep 29 21:34:19 crc kubenswrapper[4741]: E0929 21:34:19.909374 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="extract-content" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.909394 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="extract-content" Sep 29 21:34:19 crc kubenswrapper[4741]: E0929 21:34:19.909437 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="registry-server" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.909443 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="registry-server" Sep 29 21:34:19 crc kubenswrapper[4741]: E0929 21:34:19.909457 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="extract-utilities" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.909464 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="extract-utilities" Sep 29 21:34:19 crc kubenswrapper[4741]: E0929 21:34:19.909496 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19605586-361e-4c9a-9d60-685fd74ddb71" containerName="neutron-sriov-openstack-openstack-cell1" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.909502 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="19605586-361e-4c9a-9d60-685fd74ddb71" containerName="neutron-sriov-openstack-openstack-cell1" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.909738 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="19605586-361e-4c9a-9d60-685fd74ddb71" containerName="neutron-sriov-openstack-openstack-cell1" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.909767 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="23258818-8ba5-40a1-95f7-fe74390b1293" containerName="registry-server" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.910701 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.913237 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.913243 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.913679 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.913917 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.914015 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:34:19 crc kubenswrapper[4741]: I0929 21:34:19.919317 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg"] Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.011378 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-752xw\" (UniqueName: \"kubernetes.io/projected/6b5b0fa0-ac2f-4984-9585-304225817db7-kube-api-access-752xw\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.011558 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.011597 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.011628 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.011767 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.011819 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.114194 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-752xw\" (UniqueName: \"kubernetes.io/projected/6b5b0fa0-ac2f-4984-9585-304225817db7-kube-api-access-752xw\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.114644 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.114684 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.114720 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.114750 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.114778 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.120973 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.121224 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.122219 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.125892 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.129870 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.136263 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-752xw\" (UniqueName: \"kubernetes.io/projected/6b5b0fa0-ac2f-4984-9585-304225817db7-kube-api-access-752xw\") pod \"neutron-dhcp-openstack-openstack-cell1-s6xrg\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.239262 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:34:20 crc kubenswrapper[4741]: I0929 21:34:20.763176 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg"] Sep 29 21:34:21 crc kubenswrapper[4741]: I0929 21:34:21.780383 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" event={"ID":"6b5b0fa0-ac2f-4984-9585-304225817db7","Type":"ContainerStarted","Data":"e87daa3270ae734a595c4f22f0a694f96db1ce25122793d512acdb6ebb60c2c1"} Sep 29 21:34:21 crc kubenswrapper[4741]: I0929 21:34:21.780813 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" event={"ID":"6b5b0fa0-ac2f-4984-9585-304225817db7","Type":"ContainerStarted","Data":"c1ed5ffa610dcfd5f505b6853a5cf6c9ecc97154f3d27ff12a1a8999315229f9"} Sep 29 21:34:21 crc kubenswrapper[4741]: I0929 21:34:21.796303 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" podStartSLOduration=2.259955405 podStartE2EDuration="2.796285787s" podCreationTimestamp="2025-09-29 21:34:19 +0000 UTC" firstStartedPulling="2025-09-29 21:34:20.777831486 +0000 UTC m=+8702.425620818" lastFinishedPulling="2025-09-29 21:34:21.314161868 +0000 UTC m=+8702.961951200" observedRunningTime="2025-09-29 21:34:21.79609183 +0000 UTC m=+8703.443881182" watchObservedRunningTime="2025-09-29 21:34:21.796285787 +0000 UTC m=+8703.444075119" Sep 29 21:35:31 crc kubenswrapper[4741]: I0929 21:35:31.739167 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:35:31 crc kubenswrapper[4741]: I0929 21:35:31.739820 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:36:01 crc kubenswrapper[4741]: I0929 21:36:01.739068 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:36:01 crc kubenswrapper[4741]: I0929 21:36:01.739671 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:36:31 crc kubenswrapper[4741]: I0929 21:36:31.739068 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:36:31 crc kubenswrapper[4741]: I0929 21:36:31.739957 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:36:31 crc kubenswrapper[4741]: I0929 21:36:31.740019 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:36:31 crc kubenswrapper[4741]: I0929 21:36:31.740888 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:36:31 crc kubenswrapper[4741]: I0929 21:36:31.740959 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" gracePeriod=600 Sep 29 21:36:31 crc kubenswrapper[4741]: E0929 21:36:31.876610 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:36:32 crc kubenswrapper[4741]: I0929 21:36:32.047232 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" exitCode=0 Sep 29 21:36:32 crc kubenswrapper[4741]: I0929 21:36:32.047273 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2"} Sep 29 21:36:32 crc kubenswrapper[4741]: I0929 21:36:32.047305 4741 scope.go:117] "RemoveContainer" containerID="93c64ec19ecaf263112fdbcd0c747acbe7b6f772ed4bd000171f227e7e688116" Sep 29 21:36:32 crc kubenswrapper[4741]: I0929 21:36:32.048257 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:36:32 crc kubenswrapper[4741]: E0929 21:36:32.048657 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:36:45 crc kubenswrapper[4741]: I0929 21:36:45.085917 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:36:45 crc kubenswrapper[4741]: E0929 21:36:45.086679 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:36:58 crc kubenswrapper[4741]: I0929 21:36:58.086718 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:36:58 crc kubenswrapper[4741]: E0929 21:36:58.087374 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:37:13 crc kubenswrapper[4741]: I0929 21:37:13.086949 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:37:13 crc kubenswrapper[4741]: E0929 21:37:13.088329 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:37:26 crc kubenswrapper[4741]: I0929 21:37:26.086045 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:37:26 crc kubenswrapper[4741]: E0929 21:37:26.086716 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:37:38 crc kubenswrapper[4741]: I0929 21:37:38.088251 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:37:38 crc kubenswrapper[4741]: E0929 21:37:38.089065 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:37:53 crc kubenswrapper[4741]: I0929 21:37:53.086577 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:37:53 crc kubenswrapper[4741]: E0929 21:37:53.087381 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:37:58 crc kubenswrapper[4741]: I0929 21:37:58.969814 4741 generic.go:334] "Generic (PLEG): container finished" podID="6b5b0fa0-ac2f-4984-9585-304225817db7" containerID="e87daa3270ae734a595c4f22f0a694f96db1ce25122793d512acdb6ebb60c2c1" exitCode=0 Sep 29 21:37:58 crc kubenswrapper[4741]: I0929 21:37:58.969920 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" event={"ID":"6b5b0fa0-ac2f-4984-9585-304225817db7","Type":"ContainerDied","Data":"e87daa3270ae734a595c4f22f0a694f96db1ce25122793d512acdb6ebb60c2c1"} Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.426024 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.497500 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ceph\") pod \"6b5b0fa0-ac2f-4984-9585-304225817db7\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.497822 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-inventory\") pod \"6b5b0fa0-ac2f-4984-9585-304225817db7\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.497855 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ssh-key\") pod \"6b5b0fa0-ac2f-4984-9585-304225817db7\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.497902 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-752xw\" (UniqueName: \"kubernetes.io/projected/6b5b0fa0-ac2f-4984-9585-304225817db7-kube-api-access-752xw\") pod \"6b5b0fa0-ac2f-4984-9585-304225817db7\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.497959 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-agent-neutron-config-0\") pod \"6b5b0fa0-ac2f-4984-9585-304225817db7\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.498031 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-combined-ca-bundle\") pod \"6b5b0fa0-ac2f-4984-9585-304225817db7\" (UID: \"6b5b0fa0-ac2f-4984-9585-304225817db7\") " Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.504911 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5b0fa0-ac2f-4984-9585-304225817db7-kube-api-access-752xw" (OuterVolumeSpecName: "kube-api-access-752xw") pod "6b5b0fa0-ac2f-4984-9585-304225817db7" (UID: "6b5b0fa0-ac2f-4984-9585-304225817db7"). InnerVolumeSpecName "kube-api-access-752xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.505001 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ceph" (OuterVolumeSpecName: "ceph") pod "6b5b0fa0-ac2f-4984-9585-304225817db7" (UID: "6b5b0fa0-ac2f-4984-9585-304225817db7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.506708 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "6b5b0fa0-ac2f-4984-9585-304225817db7" (UID: "6b5b0fa0-ac2f-4984-9585-304225817db7"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.527647 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "6b5b0fa0-ac2f-4984-9585-304225817db7" (UID: "6b5b0fa0-ac2f-4984-9585-304225817db7"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.531291 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b5b0fa0-ac2f-4984-9585-304225817db7" (UID: "6b5b0fa0-ac2f-4984-9585-304225817db7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.549283 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-inventory" (OuterVolumeSpecName: "inventory") pod "6b5b0fa0-ac2f-4984-9585-304225817db7" (UID: "6b5b0fa0-ac2f-4984-9585-304225817db7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.600521 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.600554 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.600566 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.600579 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.600587 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-752xw\" (UniqueName: \"kubernetes.io/projected/6b5b0fa0-ac2f-4984-9585-304225817db7-kube-api-access-752xw\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.600597 4741 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b5b0fa0-ac2f-4984-9585-304225817db7-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.989595 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" event={"ID":"6b5b0fa0-ac2f-4984-9585-304225817db7","Type":"ContainerDied","Data":"c1ed5ffa610dcfd5f505b6853a5cf6c9ecc97154f3d27ff12a1a8999315229f9"} Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.989640 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1ed5ffa610dcfd5f505b6853a5cf6c9ecc97154f3d27ff12a1a8999315229f9" Sep 29 21:38:00 crc kubenswrapper[4741]: I0929 21:38:00.989693 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-s6xrg" Sep 29 21:38:04 crc kubenswrapper[4741]: I0929 21:38:04.086622 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:38:04 crc kubenswrapper[4741]: E0929 21:38:04.087955 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:38:12 crc kubenswrapper[4741]: I0929 21:38:12.240202 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 21:38:12 crc kubenswrapper[4741]: I0929 21:38:12.241273 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="e3e16528-0ced-4c87-833b-9a09283689b5" containerName="nova-cell0-conductor-conductor" containerID="cri-o://d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441" gracePeriod=30 Sep 29 21:38:12 crc kubenswrapper[4741]: I0929 21:38:12.278640 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 21:38:12 crc kubenswrapper[4741]: I0929 21:38:12.278859 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="546d78c4-996b-4f8d-af92-97982ba58890" containerName="nova-cell1-conductor-conductor" containerID="cri-o://88a52c252f6d9c79b2fc3dde07e626889fe77298cf38a87ab79861db10a63260" gracePeriod=30 Sep 29 21:38:13 crc kubenswrapper[4741]: I0929 21:38:13.979768 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 21:38:13 crc kubenswrapper[4741]: I0929 21:38:13.980296 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-log" containerID="cri-o://6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441" gracePeriod=30 Sep 29 21:38:13 crc kubenswrapper[4741]: I0929 21:38:13.980826 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-api" containerID="cri-o://5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404" gracePeriod=30 Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.046856 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.047059 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d8026026-9009-4a11-b671-d6d071a6ad7e" containerName="nova-scheduler-scheduler" containerID="cri-o://a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" gracePeriod=30 Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.119716 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.119987 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-log" containerID="cri-o://f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa" gracePeriod=30 Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.120588 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-metadata" containerID="cri-o://f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0" gracePeriod=30 Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.141468 4741 generic.go:334] "Generic (PLEG): container finished" podID="546d78c4-996b-4f8d-af92-97982ba58890" containerID="88a52c252f6d9c79b2fc3dde07e626889fe77298cf38a87ab79861db10a63260" exitCode=0 Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.141508 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"546d78c4-996b-4f8d-af92-97982ba58890","Type":"ContainerDied","Data":"88a52c252f6d9c79b2fc3dde07e626889fe77298cf38a87ab79861db10a63260"} Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.660101 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.796639 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-config-data\") pod \"546d78c4-996b-4f8d-af92-97982ba58890\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.797130 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgsjn\" (UniqueName: \"kubernetes.io/projected/546d78c4-996b-4f8d-af92-97982ba58890-kube-api-access-kgsjn\") pod \"546d78c4-996b-4f8d-af92-97982ba58890\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.797157 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-combined-ca-bundle\") pod \"546d78c4-996b-4f8d-af92-97982ba58890\" (UID: \"546d78c4-996b-4f8d-af92-97982ba58890\") " Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.806104 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/546d78c4-996b-4f8d-af92-97982ba58890-kube-api-access-kgsjn" (OuterVolumeSpecName: "kube-api-access-kgsjn") pod "546d78c4-996b-4f8d-af92-97982ba58890" (UID: "546d78c4-996b-4f8d-af92-97982ba58890"). InnerVolumeSpecName "kube-api-access-kgsjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.836785 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "546d78c4-996b-4f8d-af92-97982ba58890" (UID: "546d78c4-996b-4f8d-af92-97982ba58890"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.850154 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-config-data" (OuterVolumeSpecName: "config-data") pod "546d78c4-996b-4f8d-af92-97982ba58890" (UID: "546d78c4-996b-4f8d-af92-97982ba58890"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.901955 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgsjn\" (UniqueName: \"kubernetes.io/projected/546d78c4-996b-4f8d-af92-97982ba58890-kube-api-access-kgsjn\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.901988 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.902002 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546d78c4-996b-4f8d-af92-97982ba58890-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:14 crc kubenswrapper[4741]: I0929 21:38:14.907568 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.003530 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-config-data\") pod \"e3e16528-0ced-4c87-833b-9a09283689b5\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.003660 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-combined-ca-bundle\") pod \"e3e16528-0ced-4c87-833b-9a09283689b5\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.003845 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhbgk\" (UniqueName: \"kubernetes.io/projected/e3e16528-0ced-4c87-833b-9a09283689b5-kube-api-access-vhbgk\") pod \"e3e16528-0ced-4c87-833b-9a09283689b5\" (UID: \"e3e16528-0ced-4c87-833b-9a09283689b5\") " Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.014363 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e16528-0ced-4c87-833b-9a09283689b5-kube-api-access-vhbgk" (OuterVolumeSpecName: "kube-api-access-vhbgk") pod "e3e16528-0ced-4c87-833b-9a09283689b5" (UID: "e3e16528-0ced-4c87-833b-9a09283689b5"). InnerVolumeSpecName "kube-api-access-vhbgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.033635 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3e16528-0ced-4c87-833b-9a09283689b5" (UID: "e3e16528-0ced-4c87-833b-9a09283689b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.037459 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-config-data" (OuterVolumeSpecName: "config-data") pod "e3e16528-0ced-4c87-833b-9a09283689b5" (UID: "e3e16528-0ced-4c87-833b-9a09283689b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.086784 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.087257 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.105858 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhbgk\" (UniqueName: \"kubernetes.io/projected/e3e16528-0ced-4c87-833b-9a09283689b5-kube-api-access-vhbgk\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.105892 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.105902 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e16528-0ced-4c87-833b-9a09283689b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.114453 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.115587 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.117200 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.117233 4741 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d8026026-9009-4a11-b671-d6d071a6ad7e" containerName="nova-scheduler-scheduler" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.154674 4741 generic.go:334] "Generic (PLEG): container finished" podID="6f394e16-a1d3-45a7-af53-53898e178300" containerID="f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa" exitCode=143 Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.154754 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f394e16-a1d3-45a7-af53-53898e178300","Type":"ContainerDied","Data":"f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa"} Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.158799 4741 generic.go:334] "Generic (PLEG): container finished" podID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerID="6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441" exitCode=143 Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.158888 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ddafdfc-baf4-4368-b8c8-c3f5eafce192","Type":"ContainerDied","Data":"6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441"} Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.160624 4741 generic.go:334] "Generic (PLEG): container finished" podID="e3e16528-0ced-4c87-833b-9a09283689b5" containerID="d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441" exitCode=0 Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.160682 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e3e16528-0ced-4c87-833b-9a09283689b5","Type":"ContainerDied","Data":"d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441"} Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.160706 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e3e16528-0ced-4c87-833b-9a09283689b5","Type":"ContainerDied","Data":"c16ad00e0d4c8ab268a47839d5d0a3239bcb6612c3438221b0f0ff3e7a4ece58"} Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.160710 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.160726 4741 scope.go:117] "RemoveContainer" containerID="d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.162660 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"546d78c4-996b-4f8d-af92-97982ba58890","Type":"ContainerDied","Data":"0f272ec29b2560ae86ae8b52b7cd148e705a34ddc86bea17b5662248beaa9b07"} Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.162726 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.190597 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.193251 4741 scope.go:117] "RemoveContainer" containerID="d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441" Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.202200 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441\": container with ID starting with d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441 not found: ID does not exist" containerID="d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.202255 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441"} err="failed to get container status \"d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441\": rpc error: code = NotFound desc = could not find container \"d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441\": container with ID starting with d836a6057a392c928fad02d75919f51fdea6f876d16668458af2763a9ffdb441 not found: ID does not exist" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.202286 4741 scope.go:117] "RemoveContainer" containerID="88a52c252f6d9c79b2fc3dde07e626889fe77298cf38a87ab79861db10a63260" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.203521 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.213450 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.268081 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.276891 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.277383 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e16528-0ced-4c87-833b-9a09283689b5" containerName="nova-cell0-conductor-conductor" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.277491 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e16528-0ced-4c87-833b-9a09283689b5" containerName="nova-cell0-conductor-conductor" Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.277508 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5b0fa0-ac2f-4984-9585-304225817db7" containerName="neutron-dhcp-openstack-openstack-cell1" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.277515 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5b0fa0-ac2f-4984-9585-304225817db7" containerName="neutron-dhcp-openstack-openstack-cell1" Sep 29 21:38:15 crc kubenswrapper[4741]: E0929 21:38:15.277545 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="546d78c4-996b-4f8d-af92-97982ba58890" containerName="nova-cell1-conductor-conductor" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.277551 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="546d78c4-996b-4f8d-af92-97982ba58890" containerName="nova-cell1-conductor-conductor" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.277759 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e16528-0ced-4c87-833b-9a09283689b5" containerName="nova-cell0-conductor-conductor" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.277780 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="546d78c4-996b-4f8d-af92-97982ba58890" containerName="nova-cell1-conductor-conductor" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.277795 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5b0fa0-ac2f-4984-9585-304225817db7" containerName="neutron-dhcp-openstack-openstack-cell1" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.278582 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.282701 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.286841 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.296351 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.299517 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.302691 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.305909 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.421136 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xqm9\" (UniqueName: \"kubernetes.io/projected/ce05df11-1014-49e8-8744-3adcb02eb50b-kube-api-access-6xqm9\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.421234 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc4wc\" (UniqueName: \"kubernetes.io/projected/0bc6cc49-e74f-455c-9a7b-090cba934347-kube-api-access-lc4wc\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.421282 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce05df11-1014-49e8-8744-3adcb02eb50b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.421458 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce05df11-1014-49e8-8744-3adcb02eb50b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.421503 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc6cc49-e74f-455c-9a7b-090cba934347-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.421573 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc6cc49-e74f-455c-9a7b-090cba934347-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.523863 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce05df11-1014-49e8-8744-3adcb02eb50b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.523944 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc6cc49-e74f-455c-9a7b-090cba934347-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.524007 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc6cc49-e74f-455c-9a7b-090cba934347-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.524110 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xqm9\" (UniqueName: \"kubernetes.io/projected/ce05df11-1014-49e8-8744-3adcb02eb50b-kube-api-access-6xqm9\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.524171 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc4wc\" (UniqueName: \"kubernetes.io/projected/0bc6cc49-e74f-455c-9a7b-090cba934347-kube-api-access-lc4wc\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.524207 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce05df11-1014-49e8-8744-3adcb02eb50b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.529592 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce05df11-1014-49e8-8744-3adcb02eb50b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.539979 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce05df11-1014-49e8-8744-3adcb02eb50b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.540082 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc6cc49-e74f-455c-9a7b-090cba934347-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.540696 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc6cc49-e74f-455c-9a7b-090cba934347-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.542644 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xqm9\" (UniqueName: \"kubernetes.io/projected/ce05df11-1014-49e8-8744-3adcb02eb50b-kube-api-access-6xqm9\") pod \"nova-cell1-conductor-0\" (UID: \"ce05df11-1014-49e8-8744-3adcb02eb50b\") " pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.543473 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc4wc\" (UniqueName: \"kubernetes.io/projected/0bc6cc49-e74f-455c-9a7b-090cba934347-kube-api-access-lc4wc\") pod \"nova-cell0-conductor-0\" (UID: \"0bc6cc49-e74f-455c-9a7b-090cba934347\") " pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.599682 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:15 crc kubenswrapper[4741]: I0929 21:38:15.623522 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:16 crc kubenswrapper[4741]: I0929 21:38:16.067764 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 21:38:16 crc kubenswrapper[4741]: I0929 21:38:16.156754 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 21:38:16 crc kubenswrapper[4741]: W0929 21:38:16.162354 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce05df11_1014_49e8_8744_3adcb02eb50b.slice/crio-7faa43fc74aea3eaa2eb7e680cf1c42bd530ab1e01a9ba5a4951729f0dfb457d WatchSource:0}: Error finding container 7faa43fc74aea3eaa2eb7e680cf1c42bd530ab1e01a9ba5a4951729f0dfb457d: Status 404 returned error can't find the container with id 7faa43fc74aea3eaa2eb7e680cf1c42bd530ab1e01a9ba5a4951729f0dfb457d Sep 29 21:38:16 crc kubenswrapper[4741]: I0929 21:38:16.180436 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ce05df11-1014-49e8-8744-3adcb02eb50b","Type":"ContainerStarted","Data":"7faa43fc74aea3eaa2eb7e680cf1c42bd530ab1e01a9ba5a4951729f0dfb457d"} Sep 29 21:38:16 crc kubenswrapper[4741]: I0929 21:38:16.186754 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0bc6cc49-e74f-455c-9a7b-090cba934347","Type":"ContainerStarted","Data":"f172bbf2cdeb6aaf5ac66b6d7b27a8b3bc13571be9afc7b49b374a91f0accfeb"} Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.097656 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="546d78c4-996b-4f8d-af92-97982ba58890" path="/var/lib/kubelet/pods/546d78c4-996b-4f8d-af92-97982ba58890/volumes" Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.098729 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e16528-0ced-4c87-833b-9a09283689b5" path="/var/lib/kubelet/pods/e3e16528-0ced-4c87-833b-9a09283689b5/volumes" Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.201889 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0bc6cc49-e74f-455c-9a7b-090cba934347","Type":"ContainerStarted","Data":"589cdf8792bbedee82ba6ecfdba1505495ebad0efcacce71f8903859840e1111"} Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.202606 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.203916 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ce05df11-1014-49e8-8744-3adcb02eb50b","Type":"ContainerStarted","Data":"d21967522e4f58493c1d8bc6895e9f8d90078f88762bcce842c652f217cd5d05"} Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.204159 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.218204 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.218185686 podStartE2EDuration="2.218185686s" podCreationTimestamp="2025-09-29 21:38:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:38:17.21671472 +0000 UTC m=+8938.864504052" watchObservedRunningTime="2025-09-29 21:38:17.218185686 +0000 UTC m=+8938.865975018" Sep 29 21:38:17 crc kubenswrapper[4741]: I0929 21:38:17.239269 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.23925116 podStartE2EDuration="2.23925116s" podCreationTimestamp="2025-09-29 21:38:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:38:17.232689783 +0000 UTC m=+8938.880479125" watchObservedRunningTime="2025-09-29 21:38:17.23925116 +0000 UTC m=+8938.887040492" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.053382 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.061045 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.184878 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f394e16-a1d3-45a7-af53-53898e178300-logs\") pod \"6f394e16-a1d3-45a7-af53-53898e178300\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.185698 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f394e16-a1d3-45a7-af53-53898e178300-logs" (OuterVolumeSpecName: "logs") pod "6f394e16-a1d3-45a7-af53-53898e178300" (UID: "6f394e16-a1d3-45a7-af53-53898e178300"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.186001 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-combined-ca-bundle\") pod \"6f394e16-a1d3-45a7-af53-53898e178300\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.186030 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-config-data\") pod \"6f394e16-a1d3-45a7-af53-53898e178300\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.186088 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slggh\" (UniqueName: \"kubernetes.io/projected/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-kube-api-access-slggh\") pod \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.186104 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-logs\") pod \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.186123 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-combined-ca-bundle\") pod \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.186151 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-config-data\") pod \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\" (UID: \"6ddafdfc-baf4-4368-b8c8-c3f5eafce192\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.186169 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ljwp\" (UniqueName: \"kubernetes.io/projected/6f394e16-a1d3-45a7-af53-53898e178300-kube-api-access-8ljwp\") pod \"6f394e16-a1d3-45a7-af53-53898e178300\" (UID: \"6f394e16-a1d3-45a7-af53-53898e178300\") " Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.187004 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-logs" (OuterVolumeSpecName: "logs") pod "6ddafdfc-baf4-4368-b8c8-c3f5eafce192" (UID: "6ddafdfc-baf4-4368-b8c8-c3f5eafce192"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.188926 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-logs\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.197347 4741 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f394e16-a1d3-45a7-af53-53898e178300-logs\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.214083 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-kube-api-access-slggh" (OuterVolumeSpecName: "kube-api-access-slggh") pod "6ddafdfc-baf4-4368-b8c8-c3f5eafce192" (UID: "6ddafdfc-baf4-4368-b8c8-c3f5eafce192"). InnerVolumeSpecName "kube-api-access-slggh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.225054 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f394e16-a1d3-45a7-af53-53898e178300-kube-api-access-8ljwp" (OuterVolumeSpecName: "kube-api-access-8ljwp") pod "6f394e16-a1d3-45a7-af53-53898e178300" (UID: "6f394e16-a1d3-45a7-af53-53898e178300"). InnerVolumeSpecName "kube-api-access-8ljwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.232170 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f394e16-a1d3-45a7-af53-53898e178300" (UID: "6f394e16-a1d3-45a7-af53-53898e178300"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.253802 4741 generic.go:334] "Generic (PLEG): container finished" podID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerID="5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404" exitCode=0 Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.253870 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ddafdfc-baf4-4368-b8c8-c3f5eafce192","Type":"ContainerDied","Data":"5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404"} Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.253900 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ddafdfc-baf4-4368-b8c8-c3f5eafce192","Type":"ContainerDied","Data":"8008bb5d40ddcec41d8036c81ccfe5e19df06fc4c7c90d615ce700d7503f5696"} Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.253919 4741 scope.go:117] "RemoveContainer" containerID="5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.254071 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.255383 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-config-data" (OuterVolumeSpecName: "config-data") pod "6ddafdfc-baf4-4368-b8c8-c3f5eafce192" (UID: "6ddafdfc-baf4-4368-b8c8-c3f5eafce192"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.276646 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-config-data" (OuterVolumeSpecName: "config-data") pod "6f394e16-a1d3-45a7-af53-53898e178300" (UID: "6f394e16-a1d3-45a7-af53-53898e178300"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.279407 4741 generic.go:334] "Generic (PLEG): container finished" podID="6f394e16-a1d3-45a7-af53-53898e178300" containerID="f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0" exitCode=0 Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.281039 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.286779 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f394e16-a1d3-45a7-af53-53898e178300","Type":"ContainerDied","Data":"f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0"} Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.286989 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f394e16-a1d3-45a7-af53-53898e178300","Type":"ContainerDied","Data":"93086df0952d7c43458a361aad0842c2bcf513681c8efa5677cbfa2993103c39"} Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.299477 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.299508 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f394e16-a1d3-45a7-af53-53898e178300-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.299520 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slggh\" (UniqueName: \"kubernetes.io/projected/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-kube-api-access-slggh\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.299547 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.299559 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ljwp\" (UniqueName: \"kubernetes.io/projected/6f394e16-a1d3-45a7-af53-53898e178300-kube-api-access-8ljwp\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.324881 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ddafdfc-baf4-4368-b8c8-c3f5eafce192" (UID: "6ddafdfc-baf4-4368-b8c8-c3f5eafce192"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.397358 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f394e16_a1d3_45a7_af53_53898e178300.slice/crio-93086df0952d7c43458a361aad0842c2bcf513681c8efa5677cbfa2993103c39\": RecentStats: unable to find data in memory cache]" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.403822 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ddafdfc-baf4-4368-b8c8-c3f5eafce192-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.444162 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.449961 4741 scope.go:117] "RemoveContainer" containerID="6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.464689 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.474491 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.475259 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-log" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.475382 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-log" Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.475549 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-metadata" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.475622 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-metadata" Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.475693 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-api" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.475761 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-api" Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.475896 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-log" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.475980 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-log" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.476371 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-api" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.476503 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-log" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.476588 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f394e16-a1d3-45a7-af53-53898e178300" containerName="nova-metadata-metadata" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.476672 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" containerName="nova-api-log" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.478556 4741 scope.go:117] "RemoveContainer" containerID="5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.478961 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.482691 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.487734 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404\": container with ID starting with 5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404 not found: ID does not exist" containerID="5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.487778 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404"} err="failed to get container status \"5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404\": rpc error: code = NotFound desc = could not find container \"5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404\": container with ID starting with 5fe6fce998d856174b0ad5e6f61f83efd797e7aed83649c7301eb899355c0404 not found: ID does not exist" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.487810 4741 scope.go:117] "RemoveContainer" containerID="6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.489625 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.490525 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441\": container with ID starting with 6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441 not found: ID does not exist" containerID="6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.490586 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441"} err="failed to get container status \"6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441\": rpc error: code = NotFound desc = could not find container \"6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441\": container with ID starting with 6ffe6a845c00753bb54e4dfdd96c61fd5c420d201ffe5d5940b82b3e7c8a7441 not found: ID does not exist" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.490615 4741 scope.go:117] "RemoveContainer" containerID="f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.514218 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa65212c-f217-460f-adda-71813980fb4c-config-data\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.514799 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa65212c-f217-460f-adda-71813980fb4c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.515112 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa65212c-f217-460f-adda-71813980fb4c-logs\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.515275 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2q5x\" (UniqueName: \"kubernetes.io/projected/aa65212c-f217-460f-adda-71813980fb4c-kube-api-access-b2q5x\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.555670 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8"] Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.565200 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.569243 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8"] Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.570869 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.570926 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.570883 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-gn42d" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.570878 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.571175 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.571210 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.571271 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.581258 4741 scope.go:117] "RemoveContainer" containerID="f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.618913 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa65212c-f217-460f-adda-71813980fb4c-config-data\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.619030 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa65212c-f217-460f-adda-71813980fb4c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.619082 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa65212c-f217-460f-adda-71813980fb4c-logs\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.619113 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2q5x\" (UniqueName: \"kubernetes.io/projected/aa65212c-f217-460f-adda-71813980fb4c-kube-api-access-b2q5x\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.620679 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa65212c-f217-460f-adda-71813980fb4c-logs\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.628913 4741 scope.go:117] "RemoveContainer" containerID="f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.629159 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.631339 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa65212c-f217-460f-adda-71813980fb4c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.631841 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa65212c-f217-460f-adda-71813980fb4c-config-data\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.632245 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0\": container with ID starting with f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0 not found: ID does not exist" containerID="f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.636439 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0"} err="failed to get container status \"f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0\": rpc error: code = NotFound desc = could not find container \"f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0\": container with ID starting with f5836cc37a691ec07c83df2f6108bfdec1fa8bd42df19367cab243b46bc7c2a0 not found: ID does not exist" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.640488 4741 scope.go:117] "RemoveContainer" containerID="f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.642629 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: E0929 21:38:18.642633 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa\": container with ID starting with f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa not found: ID does not exist" containerID="f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.642820 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa"} err="failed to get container status \"f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa\": rpc error: code = NotFound desc = could not find container \"f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa\": container with ID starting with f776bdf4d39be8af1443e526467c66f6e836761c093b7f473c0d2ee9784092fa not found: ID does not exist" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.645866 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2q5x\" (UniqueName: \"kubernetes.io/projected/aa65212c-f217-460f-adda-71813980fb4c-kube-api-access-b2q5x\") pod \"nova-metadata-0\" (UID: \"aa65212c-f217-460f-adda-71813980fb4c\") " pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.657924 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.659981 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.663421 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.684125 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.720675 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.720732 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.720755 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.721834 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxzbc\" (UniqueName: \"kubernetes.io/projected/b6c8b549-a934-4c9d-919c-b5dee30d4385-kube-api-access-mxzbc\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.721885 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.721952 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.722035 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.722212 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.722287 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.722520 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.722640 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.817454 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.838854 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.838946 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.838983 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0863919b-b3fc-40b0-9957-8cc9573c2e3b-logs\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839011 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839044 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839167 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839223 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839373 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxzbc\" (UniqueName: \"kubernetes.io/projected/b6c8b549-a934-4c9d-919c-b5dee30d4385-kube-api-access-mxzbc\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839516 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839601 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839699 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0863919b-b3fc-40b0-9957-8cc9573c2e3b-config-data\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839770 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839794 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx9dz\" (UniqueName: \"kubernetes.io/projected/0863919b-b3fc-40b0-9957-8cc9573c2e3b-kube-api-access-gx9dz\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839841 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.839917 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0863919b-b3fc-40b0-9957-8cc9573c2e3b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.841019 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.841663 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.846805 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.848781 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.850218 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.850451 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.853087 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.858027 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.860461 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.860475 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.861126 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxzbc\" (UniqueName: \"kubernetes.io/projected/b6c8b549-a934-4c9d-919c-b5dee30d4385-kube-api-access-mxzbc\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.887989 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.941632 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0863919b-b3fc-40b0-9957-8cc9573c2e3b-logs\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.941766 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0863919b-b3fc-40b0-9957-8cc9573c2e3b-config-data\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.941795 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx9dz\" (UniqueName: \"kubernetes.io/projected/0863919b-b3fc-40b0-9957-8cc9573c2e3b-kube-api-access-gx9dz\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.941833 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0863919b-b3fc-40b0-9957-8cc9573c2e3b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.942907 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0863919b-b3fc-40b0-9957-8cc9573c2e3b-logs\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.945313 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0863919b-b3fc-40b0-9957-8cc9573c2e3b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.952049 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0863919b-b3fc-40b0-9957-8cc9573c2e3b-config-data\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:18 crc kubenswrapper[4741]: I0929 21:38:18.963391 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx9dz\" (UniqueName: \"kubernetes.io/projected/0863919b-b3fc-40b0-9957-8cc9573c2e3b-kube-api-access-gx9dz\") pod \"nova-api-0\" (UID: \"0863919b-b3fc-40b0-9957-8cc9573c2e3b\") " pod="openstack/nova-api-0" Sep 29 21:38:19 crc kubenswrapper[4741]: I0929 21:38:19.002915 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 21:38:19 crc kubenswrapper[4741]: I0929 21:38:19.131954 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ddafdfc-baf4-4368-b8c8-c3f5eafce192" path="/var/lib/kubelet/pods/6ddafdfc-baf4-4368-b8c8-c3f5eafce192/volumes" Sep 29 21:38:19 crc kubenswrapper[4741]: I0929 21:38:19.145993 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f394e16-a1d3-45a7-af53-53898e178300" path="/var/lib/kubelet/pods/6f394e16-a1d3-45a7-af53-53898e178300/volumes" Sep 29 21:38:19 crc kubenswrapper[4741]: I0929 21:38:19.487611 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 21:38:19 crc kubenswrapper[4741]: W0929 21:38:19.488691 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa65212c_f217_460f_adda_71813980fb4c.slice/crio-f6c4d8068cc6f8afd84b8e550a28706662dcc7801c6923e7cc7b42c2080444b6 WatchSource:0}: Error finding container f6c4d8068cc6f8afd84b8e550a28706662dcc7801c6923e7cc7b42c2080444b6: Status 404 returned error can't find the container with id f6c4d8068cc6f8afd84b8e550a28706662dcc7801c6923e7cc7b42c2080444b6 Sep 29 21:38:19 crc kubenswrapper[4741]: I0929 21:38:19.661331 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 21:38:19 crc kubenswrapper[4741]: I0929 21:38:19.764478 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8"] Sep 29 21:38:19 crc kubenswrapper[4741]: I0929 21:38:19.785064 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:38:20 crc kubenswrapper[4741]: E0929 21:38:20.112087 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23 is running failed: container process not found" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 21:38:20 crc kubenswrapper[4741]: E0929 21:38:20.112610 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23 is running failed: container process not found" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 21:38:20 crc kubenswrapper[4741]: E0929 21:38:20.113023 4741 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23 is running failed: container process not found" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 21:38:20 crc kubenswrapper[4741]: E0929 21:38:20.113064 4741 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d8026026-9009-4a11-b671-d6d071a6ad7e" containerName="nova-scheduler-scheduler" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.154266 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.262066 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.263545 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zd7s\" (UniqueName: \"kubernetes.io/projected/d8026026-9009-4a11-b671-d6d071a6ad7e-kube-api-access-8zd7s\") pod \"d8026026-9009-4a11-b671-d6d071a6ad7e\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.263586 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-combined-ca-bundle\") pod \"d8026026-9009-4a11-b671-d6d071a6ad7e\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.263902 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-config-data\") pod \"d8026026-9009-4a11-b671-d6d071a6ad7e\" (UID: \"d8026026-9009-4a11-b671-d6d071a6ad7e\") " Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.271890 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8026026-9009-4a11-b671-d6d071a6ad7e-kube-api-access-8zd7s" (OuterVolumeSpecName: "kube-api-access-8zd7s") pod "d8026026-9009-4a11-b671-d6d071a6ad7e" (UID: "d8026026-9009-4a11-b671-d6d071a6ad7e"). InnerVolumeSpecName "kube-api-access-8zd7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.337220 4741 generic.go:334] "Generic (PLEG): container finished" podID="d8026026-9009-4a11-b671-d6d071a6ad7e" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" exitCode=0 Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.337321 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.337561 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d8026026-9009-4a11-b671-d6d071a6ad7e","Type":"ContainerDied","Data":"a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.337587 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d8026026-9009-4a11-b671-d6d071a6ad7e","Type":"ContainerDied","Data":"2d10164788c3b18d922d1f9725776e3d7e60fe3b29d9c177a49e361c71125fb4"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.337602 4741 scope.go:117] "RemoveContainer" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.343881 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0863919b-b3fc-40b0-9957-8cc9573c2e3b","Type":"ContainerStarted","Data":"7ea686a8738243ace841523f5f9ddb4f20d84cf24fa0c04e3f33799b55be9eb1"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.343914 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0863919b-b3fc-40b0-9957-8cc9573c2e3b","Type":"ContainerStarted","Data":"7354c37cd2f4e1b1a032d3ce0463ddd8a4be2a1abaf2dfc5cc522fe780c527ed"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.347030 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa65212c-f217-460f-adda-71813980fb4c","Type":"ContainerStarted","Data":"1aea3c0dbf552f1c4333a7a1352c65ecd379b4be2ec7a3dbc4ea47c03c38b4c5"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.347076 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa65212c-f217-460f-adda-71813980fb4c","Type":"ContainerStarted","Data":"94f31e1b2204825509eb068c7ef170623e0ef6e48355d594c3982a2a646673d9"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.347091 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa65212c-f217-460f-adda-71813980fb4c","Type":"ContainerStarted","Data":"f6c4d8068cc6f8afd84b8e550a28706662dcc7801c6923e7cc7b42c2080444b6"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.349162 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" event={"ID":"b6c8b549-a934-4c9d-919c-b5dee30d4385","Type":"ContainerStarted","Data":"b9aacf87b00118ac4a98a8a8f4a51ba5c84a4b184ea1ff775c35e9730ac84d47"} Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.362923 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8026026-9009-4a11-b671-d6d071a6ad7e" (UID: "d8026026-9009-4a11-b671-d6d071a6ad7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.367999 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zd7s\" (UniqueName: \"kubernetes.io/projected/d8026026-9009-4a11-b671-d6d071a6ad7e-kube-api-access-8zd7s\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.368025 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.378615 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-config-data" (OuterVolumeSpecName: "config-data") pod "d8026026-9009-4a11-b671-d6d071a6ad7e" (UID: "d8026026-9009-4a11-b671-d6d071a6ad7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.391828 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.391807863 podStartE2EDuration="2.391807863s" podCreationTimestamp="2025-09-29 21:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:38:20.363865513 +0000 UTC m=+8942.011654865" watchObservedRunningTime="2025-09-29 21:38:20.391807863 +0000 UTC m=+8942.039597195" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.472234 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8026026-9009-4a11-b671-d6d071a6ad7e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.522650 4741 scope.go:117] "RemoveContainer" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" Sep 29 21:38:20 crc kubenswrapper[4741]: E0929 21:38:20.523517 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23\": container with ID starting with a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23 not found: ID does not exist" containerID="a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.523544 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23"} err="failed to get container status \"a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23\": rpc error: code = NotFound desc = could not find container \"a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23\": container with ID starting with a7cfa9509f698769a27b26b734a7f8b9e87954f45630defe7a1a207b67fb6b23 not found: ID does not exist" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.672475 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.682535 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.695512 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 21:38:20 crc kubenswrapper[4741]: E0929 21:38:20.696101 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8026026-9009-4a11-b671-d6d071a6ad7e" containerName="nova-scheduler-scheduler" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.696125 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8026026-9009-4a11-b671-d6d071a6ad7e" containerName="nova-scheduler-scheduler" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.696374 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8026026-9009-4a11-b671-d6d071a6ad7e" containerName="nova-scheduler-scheduler" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.697450 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.708223 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.720229 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.776886 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-865qr\" (UniqueName: \"kubernetes.io/projected/001b7218-494b-4004-ad8c-4474aabad9dd-kube-api-access-865qr\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.777021 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001b7218-494b-4004-ad8c-4474aabad9dd-config-data\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.777048 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001b7218-494b-4004-ad8c-4474aabad9dd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.878062 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-865qr\" (UniqueName: \"kubernetes.io/projected/001b7218-494b-4004-ad8c-4474aabad9dd-kube-api-access-865qr\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.878191 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001b7218-494b-4004-ad8c-4474aabad9dd-config-data\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.878239 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001b7218-494b-4004-ad8c-4474aabad9dd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.882834 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001b7218-494b-4004-ad8c-4474aabad9dd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.885934 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001b7218-494b-4004-ad8c-4474aabad9dd-config-data\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:20 crc kubenswrapper[4741]: I0929 21:38:20.903409 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-865qr\" (UniqueName: \"kubernetes.io/projected/001b7218-494b-4004-ad8c-4474aabad9dd-kube-api-access-865qr\") pod \"nova-scheduler-0\" (UID: \"001b7218-494b-4004-ad8c-4474aabad9dd\") " pod="openstack/nova-scheduler-0" Sep 29 21:38:21 crc kubenswrapper[4741]: I0929 21:38:21.016733 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 21:38:21 crc kubenswrapper[4741]: I0929 21:38:21.131999 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8026026-9009-4a11-b671-d6d071a6ad7e" path="/var/lib/kubelet/pods/d8026026-9009-4a11-b671-d6d071a6ad7e/volumes" Sep 29 21:38:21 crc kubenswrapper[4741]: I0929 21:38:21.358806 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" event={"ID":"b6c8b549-a934-4c9d-919c-b5dee30d4385","Type":"ContainerStarted","Data":"b927dc65d5526a3b4c8678a09a011b8e934c4458c0c6f854ce6018f56e465ac8"} Sep 29 21:38:21 crc kubenswrapper[4741]: I0929 21:38:21.361803 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0863919b-b3fc-40b0-9957-8cc9573c2e3b","Type":"ContainerStarted","Data":"e495d38febb62a54475828c0f996def0ced798b7100668d58899607ed71cf794"} Sep 29 21:38:21 crc kubenswrapper[4741]: I0929 21:38:21.386611 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" podStartSLOduration=2.9122582809999997 podStartE2EDuration="3.386595099s" podCreationTimestamp="2025-09-29 21:38:18 +0000 UTC" firstStartedPulling="2025-09-29 21:38:19.784877812 +0000 UTC m=+8941.432667154" lastFinishedPulling="2025-09-29 21:38:20.25921464 +0000 UTC m=+8941.907003972" observedRunningTime="2025-09-29 21:38:21.377380359 +0000 UTC m=+8943.025169711" watchObservedRunningTime="2025-09-29 21:38:21.386595099 +0000 UTC m=+8943.034384431" Sep 29 21:38:21 crc kubenswrapper[4741]: I0929 21:38:21.401846 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.401820958 podStartE2EDuration="3.401820958s" podCreationTimestamp="2025-09-29 21:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:38:21.397350868 +0000 UTC m=+8943.045140200" watchObservedRunningTime="2025-09-29 21:38:21.401820958 +0000 UTC m=+8943.049610290" Sep 29 21:38:21 crc kubenswrapper[4741]: I0929 21:38:21.484743 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 21:38:22 crc kubenswrapper[4741]: I0929 21:38:22.371766 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"001b7218-494b-4004-ad8c-4474aabad9dd","Type":"ContainerStarted","Data":"7bf645efe869014a8462bb084607e689537441e5e5be4250e98739cd2b7d1d8a"} Sep 29 21:38:22 crc kubenswrapper[4741]: I0929 21:38:22.372266 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"001b7218-494b-4004-ad8c-4474aabad9dd","Type":"ContainerStarted","Data":"25d33db0679ccfcbb2fb978d16188b1927701055d292ea6914a8760ea14e5d7e"} Sep 29 21:38:22 crc kubenswrapper[4741]: I0929 21:38:22.391488 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.391468494 podStartE2EDuration="2.391468494s" podCreationTimestamp="2025-09-29 21:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 21:38:22.390282686 +0000 UTC m=+8944.038072038" watchObservedRunningTime="2025-09-29 21:38:22.391468494 +0000 UTC m=+8944.039257826" Sep 29 21:38:23 crc kubenswrapper[4741]: I0929 21:38:23.817745 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 21:38:23 crc kubenswrapper[4741]: I0929 21:38:23.818522 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 21:38:25 crc kubenswrapper[4741]: I0929 21:38:25.634418 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 29 21:38:25 crc kubenswrapper[4741]: I0929 21:38:25.658638 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 29 21:38:26 crc kubenswrapper[4741]: I0929 21:38:26.017541 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 21:38:28 crc kubenswrapper[4741]: I0929 21:38:28.818012 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 21:38:28 crc kubenswrapper[4741]: I0929 21:38:28.818627 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 21:38:29 crc kubenswrapper[4741]: I0929 21:38:29.003344 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 21:38:29 crc kubenswrapper[4741]: I0929 21:38:29.003426 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 21:38:29 crc kubenswrapper[4741]: I0929 21:38:29.901584 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="aa65212c-f217-460f-adda-71813980fb4c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.188:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 21:38:29 crc kubenswrapper[4741]: I0929 21:38:29.901917 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="aa65212c-f217-460f-adda-71813980fb4c" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.188:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 21:38:30 crc kubenswrapper[4741]: I0929 21:38:30.085573 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0863919b-b3fc-40b0-9957-8cc9573c2e3b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 21:38:30 crc kubenswrapper[4741]: I0929 21:38:30.085991 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:38:30 crc kubenswrapper[4741]: I0929 21:38:30.085574 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0863919b-b3fc-40b0-9957-8cc9573c2e3b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 21:38:30 crc kubenswrapper[4741]: E0929 21:38:30.086278 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:38:31 crc kubenswrapper[4741]: I0929 21:38:31.016528 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 21:38:31 crc kubenswrapper[4741]: I0929 21:38:31.611020 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 21:38:31 crc kubenswrapper[4741]: I0929 21:38:31.644625 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 21:38:38 crc kubenswrapper[4741]: I0929 21:38:38.819693 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 21:38:38 crc kubenswrapper[4741]: I0929 21:38:38.820384 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 21:38:38 crc kubenswrapper[4741]: I0929 21:38:38.823173 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 21:38:38 crc kubenswrapper[4741]: I0929 21:38:38.823967 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 21:38:39 crc kubenswrapper[4741]: I0929 21:38:39.011461 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 21:38:39 crc kubenswrapper[4741]: I0929 21:38:39.012654 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 21:38:39 crc kubenswrapper[4741]: I0929 21:38:39.013085 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 21:38:39 crc kubenswrapper[4741]: I0929 21:38:39.013117 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 21:38:39 crc kubenswrapper[4741]: I0929 21:38:39.019667 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 21:38:39 crc kubenswrapper[4741]: I0929 21:38:39.033694 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 21:38:45 crc kubenswrapper[4741]: I0929 21:38:45.086251 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:38:45 crc kubenswrapper[4741]: E0929 21:38:45.087222 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:38:56 crc kubenswrapper[4741]: I0929 21:38:56.087022 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:38:56 crc kubenswrapper[4741]: E0929 21:38:56.088196 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:39:11 crc kubenswrapper[4741]: I0929 21:39:11.085859 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:39:11 crc kubenswrapper[4741]: E0929 21:39:11.086721 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:39:26 crc kubenswrapper[4741]: I0929 21:39:26.086669 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:39:26 crc kubenswrapper[4741]: E0929 21:39:26.089431 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:39:37 crc kubenswrapper[4741]: I0929 21:39:37.087103 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:39:37 crc kubenswrapper[4741]: E0929 21:39:37.087979 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:39:52 crc kubenswrapper[4741]: I0929 21:39:52.086457 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:39:52 crc kubenswrapper[4741]: E0929 21:39:52.087190 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:40:04 crc kubenswrapper[4741]: I0929 21:40:04.086490 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:40:04 crc kubenswrapper[4741]: E0929 21:40:04.087270 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:40:19 crc kubenswrapper[4741]: I0929 21:40:19.094316 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:40:19 crc kubenswrapper[4741]: E0929 21:40:19.095164 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:40:31 crc kubenswrapper[4741]: I0929 21:40:31.086147 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:40:31 crc kubenswrapper[4741]: E0929 21:40:31.086928 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:40:46 crc kubenswrapper[4741]: I0929 21:40:46.085823 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:40:46 crc kubenswrapper[4741]: E0929 21:40:46.086620 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:41:00 crc kubenswrapper[4741]: I0929 21:41:00.086205 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:41:00 crc kubenswrapper[4741]: E0929 21:41:00.086986 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:41:11 crc kubenswrapper[4741]: I0929 21:41:11.086882 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:41:11 crc kubenswrapper[4741]: E0929 21:41:11.087808 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:41:25 crc kubenswrapper[4741]: I0929 21:41:25.088484 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:41:25 crc kubenswrapper[4741]: E0929 21:41:25.090836 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:41:40 crc kubenswrapper[4741]: I0929 21:41:40.094869 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:41:41 crc kubenswrapper[4741]: I0929 21:41:41.348186 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"3a12d2eb3bbc17d7730f91f2eaa48266afea0ad8837e02e3139eea4a70466b1a"} Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.209430 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k58qt"] Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.212076 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.226391 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k58qt"] Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.284827 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws4kw\" (UniqueName: \"kubernetes.io/projected/e37df100-0174-4b51-9134-3f63009006c5-kube-api-access-ws4kw\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.286698 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-utilities\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.287047 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-catalog-content\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.389836 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-utilities\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.389916 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-catalog-content\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.389986 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws4kw\" (UniqueName: \"kubernetes.io/projected/e37df100-0174-4b51-9134-3f63009006c5-kube-api-access-ws4kw\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.390896 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-catalog-content\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.390906 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-utilities\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.414123 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws4kw\" (UniqueName: \"kubernetes.io/projected/e37df100-0174-4b51-9134-3f63009006c5-kube-api-access-ws4kw\") pod \"redhat-marketplace-k58qt\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.536512 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:26 crc kubenswrapper[4741]: I0929 21:42:26.989925 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k58qt"] Sep 29 21:42:27 crc kubenswrapper[4741]: I0929 21:42:27.785602 4741 generic.go:334] "Generic (PLEG): container finished" podID="e37df100-0174-4b51-9134-3f63009006c5" containerID="93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9" exitCode=0 Sep 29 21:42:27 crc kubenswrapper[4741]: I0929 21:42:27.785647 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k58qt" event={"ID":"e37df100-0174-4b51-9134-3f63009006c5","Type":"ContainerDied","Data":"93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9"} Sep 29 21:42:27 crc kubenswrapper[4741]: I0929 21:42:27.785673 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k58qt" event={"ID":"e37df100-0174-4b51-9134-3f63009006c5","Type":"ContainerStarted","Data":"6684ebb28014a170b2a9f103d5fd73da8c15f1f8b42b738c091cc52942f7b8d0"} Sep 29 21:42:29 crc kubenswrapper[4741]: I0929 21:42:29.807020 4741 generic.go:334] "Generic (PLEG): container finished" podID="e37df100-0174-4b51-9134-3f63009006c5" containerID="f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446" exitCode=0 Sep 29 21:42:29 crc kubenswrapper[4741]: I0929 21:42:29.807100 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k58qt" event={"ID":"e37df100-0174-4b51-9134-3f63009006c5","Type":"ContainerDied","Data":"f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446"} Sep 29 21:42:30 crc kubenswrapper[4741]: I0929 21:42:30.824202 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k58qt" event={"ID":"e37df100-0174-4b51-9134-3f63009006c5","Type":"ContainerStarted","Data":"22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763"} Sep 29 21:42:30 crc kubenswrapper[4741]: I0929 21:42:30.845518 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k58qt" podStartSLOduration=2.418621357 podStartE2EDuration="4.845497272s" podCreationTimestamp="2025-09-29 21:42:26 +0000 UTC" firstStartedPulling="2025-09-29 21:42:27.787748483 +0000 UTC m=+9189.435537815" lastFinishedPulling="2025-09-29 21:42:30.214624398 +0000 UTC m=+9191.862413730" observedRunningTime="2025-09-29 21:42:30.843251611 +0000 UTC m=+9192.491040953" watchObservedRunningTime="2025-09-29 21:42:30.845497272 +0000 UTC m=+9192.493286604" Sep 29 21:42:36 crc kubenswrapper[4741]: I0929 21:42:36.536955 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:36 crc kubenswrapper[4741]: I0929 21:42:36.537601 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:36 crc kubenswrapper[4741]: I0929 21:42:36.594136 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:36 crc kubenswrapper[4741]: I0929 21:42:36.935577 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:36 crc kubenswrapper[4741]: I0929 21:42:36.991787 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k58qt"] Sep 29 21:42:38 crc kubenswrapper[4741]: I0929 21:42:38.903261 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k58qt" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="registry-server" containerID="cri-o://22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763" gracePeriod=2 Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.426561 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.504935 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws4kw\" (UniqueName: \"kubernetes.io/projected/e37df100-0174-4b51-9134-3f63009006c5-kube-api-access-ws4kw\") pod \"e37df100-0174-4b51-9134-3f63009006c5\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.505058 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-catalog-content\") pod \"e37df100-0174-4b51-9134-3f63009006c5\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.505182 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-utilities\") pod \"e37df100-0174-4b51-9134-3f63009006c5\" (UID: \"e37df100-0174-4b51-9134-3f63009006c5\") " Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.506588 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-utilities" (OuterVolumeSpecName: "utilities") pod "e37df100-0174-4b51-9134-3f63009006c5" (UID: "e37df100-0174-4b51-9134-3f63009006c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.512085 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37df100-0174-4b51-9134-3f63009006c5-kube-api-access-ws4kw" (OuterVolumeSpecName: "kube-api-access-ws4kw") pod "e37df100-0174-4b51-9134-3f63009006c5" (UID: "e37df100-0174-4b51-9134-3f63009006c5"). InnerVolumeSpecName "kube-api-access-ws4kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.525886 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e37df100-0174-4b51-9134-3f63009006c5" (UID: "e37df100-0174-4b51-9134-3f63009006c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.607862 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.607913 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37df100-0174-4b51-9134-3f63009006c5-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.607931 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws4kw\" (UniqueName: \"kubernetes.io/projected/e37df100-0174-4b51-9134-3f63009006c5-kube-api-access-ws4kw\") on node \"crc\" DevicePath \"\"" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.916169 4741 generic.go:334] "Generic (PLEG): container finished" podID="e37df100-0174-4b51-9134-3f63009006c5" containerID="22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763" exitCode=0 Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.916207 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k58qt" event={"ID":"e37df100-0174-4b51-9134-3f63009006c5","Type":"ContainerDied","Data":"22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763"} Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.916232 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k58qt" event={"ID":"e37df100-0174-4b51-9134-3f63009006c5","Type":"ContainerDied","Data":"6684ebb28014a170b2a9f103d5fd73da8c15f1f8b42b738c091cc52942f7b8d0"} Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.916248 4741 scope.go:117] "RemoveContainer" containerID="22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.916347 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k58qt" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.955987 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k58qt"] Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.956649 4741 scope.go:117] "RemoveContainer" containerID="f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446" Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.986220 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k58qt"] Sep 29 21:42:39 crc kubenswrapper[4741]: I0929 21:42:39.987132 4741 scope.go:117] "RemoveContainer" containerID="93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9" Sep 29 21:42:40 crc kubenswrapper[4741]: I0929 21:42:40.033164 4741 scope.go:117] "RemoveContainer" containerID="22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763" Sep 29 21:42:40 crc kubenswrapper[4741]: E0929 21:42:40.036047 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763\": container with ID starting with 22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763 not found: ID does not exist" containerID="22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763" Sep 29 21:42:40 crc kubenswrapper[4741]: I0929 21:42:40.036090 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763"} err="failed to get container status \"22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763\": rpc error: code = NotFound desc = could not find container \"22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763\": container with ID starting with 22a05cbb39305d22e4b803b91b0b0c105e1b0635caf89d81d015a1e7f1dae763 not found: ID does not exist" Sep 29 21:42:40 crc kubenswrapper[4741]: I0929 21:42:40.036117 4741 scope.go:117] "RemoveContainer" containerID="f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446" Sep 29 21:42:40 crc kubenswrapper[4741]: E0929 21:42:40.036428 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446\": container with ID starting with f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446 not found: ID does not exist" containerID="f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446" Sep 29 21:42:40 crc kubenswrapper[4741]: I0929 21:42:40.036482 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446"} err="failed to get container status \"f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446\": rpc error: code = NotFound desc = could not find container \"f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446\": container with ID starting with f78f65665a870b4150e93f9004c2b3676aea3ef44a5b38704c5152d87b291446 not found: ID does not exist" Sep 29 21:42:40 crc kubenswrapper[4741]: I0929 21:42:40.036518 4741 scope.go:117] "RemoveContainer" containerID="93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9" Sep 29 21:42:40 crc kubenswrapper[4741]: E0929 21:42:40.037184 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9\": container with ID starting with 93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9 not found: ID does not exist" containerID="93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9" Sep 29 21:42:40 crc kubenswrapper[4741]: I0929 21:42:40.037217 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9"} err="failed to get container status \"93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9\": rpc error: code = NotFound desc = could not find container \"93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9\": container with ID starting with 93047adc98189b39616a67ba215a4cb27781d1cf250d52a7aa5653dcdcc21ef9 not found: ID does not exist" Sep 29 21:42:41 crc kubenswrapper[4741]: I0929 21:42:41.097783 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e37df100-0174-4b51-9134-3f63009006c5" path="/var/lib/kubelet/pods/e37df100-0174-4b51-9134-3f63009006c5/volumes" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.117850 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-85nhj"] Sep 29 21:43:30 crc kubenswrapper[4741]: E0929 21:43:30.119093 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="extract-utilities" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.119111 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="extract-utilities" Sep 29 21:43:30 crc kubenswrapper[4741]: E0929 21:43:30.119125 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="extract-content" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.119131 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="extract-content" Sep 29 21:43:30 crc kubenswrapper[4741]: E0929 21:43:30.119157 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="registry-server" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.119162 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="registry-server" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.119427 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37df100-0174-4b51-9134-3f63009006c5" containerName="registry-server" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.121555 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.128123 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85nhj"] Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.242913 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq72t\" (UniqueName: \"kubernetes.io/projected/32dfad7c-e541-4a01-a3d0-2520ffd05128-kube-api-access-cq72t\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.242991 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-catalog-content\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.243177 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-utilities\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.320059 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dvp42"] Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.322917 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.335650 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dvp42"] Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.346254 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq72t\" (UniqueName: \"kubernetes.io/projected/32dfad7c-e541-4a01-a3d0-2520ffd05128-kube-api-access-cq72t\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.346340 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-catalog-content\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.346488 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-utilities\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.347237 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-utilities\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.353812 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-catalog-content\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.380413 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq72t\" (UniqueName: \"kubernetes.io/projected/32dfad7c-e541-4a01-a3d0-2520ffd05128-kube-api-access-cq72t\") pod \"redhat-operators-85nhj\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.443399 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.448821 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-catalog-content\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.448920 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-utilities\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.449211 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4js5t\" (UniqueName: \"kubernetes.io/projected/719f915f-590f-4363-a719-686c7dc3f9fb-kube-api-access-4js5t\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.551897 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4js5t\" (UniqueName: \"kubernetes.io/projected/719f915f-590f-4363-a719-686c7dc3f9fb-kube-api-access-4js5t\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.552484 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-catalog-content\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.552525 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-utilities\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.552981 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-utilities\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.553103 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-catalog-content\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.578433 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4js5t\" (UniqueName: \"kubernetes.io/projected/719f915f-590f-4363-a719-686c7dc3f9fb-kube-api-access-4js5t\") pod \"certified-operators-dvp42\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:30 crc kubenswrapper[4741]: I0929 21:43:30.644824 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:31 crc kubenswrapper[4741]: I0929 21:43:31.053739 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85nhj"] Sep 29 21:43:31 crc kubenswrapper[4741]: I0929 21:43:31.296887 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dvp42"] Sep 29 21:43:31 crc kubenswrapper[4741]: W0929 21:43:31.303678 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod719f915f_590f_4363_a719_686c7dc3f9fb.slice/crio-e2387587ee1c51a4d5ee9d2018323601f9f3df2af0275c393d7f8044c1771e53 WatchSource:0}: Error finding container e2387587ee1c51a4d5ee9d2018323601f9f3df2af0275c393d7f8044c1771e53: Status 404 returned error can't find the container with id e2387587ee1c51a4d5ee9d2018323601f9f3df2af0275c393d7f8044c1771e53 Sep 29 21:43:31 crc kubenswrapper[4741]: I0929 21:43:31.438598 4741 generic.go:334] "Generic (PLEG): container finished" podID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerID="58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e" exitCode=0 Sep 29 21:43:31 crc kubenswrapper[4741]: I0929 21:43:31.438707 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85nhj" event={"ID":"32dfad7c-e541-4a01-a3d0-2520ffd05128","Type":"ContainerDied","Data":"58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e"} Sep 29 21:43:31 crc kubenswrapper[4741]: I0929 21:43:31.439200 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85nhj" event={"ID":"32dfad7c-e541-4a01-a3d0-2520ffd05128","Type":"ContainerStarted","Data":"df0554b1429657f1e472d26049a94557cfdb72780c03e78785875ea25fdfbd05"} Sep 29 21:43:31 crc kubenswrapper[4741]: I0929 21:43:31.444646 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvp42" event={"ID":"719f915f-590f-4363-a719-686c7dc3f9fb","Type":"ContainerStarted","Data":"e2387587ee1c51a4d5ee9d2018323601f9f3df2af0275c393d7f8044c1771e53"} Sep 29 21:43:31 crc kubenswrapper[4741]: I0929 21:43:31.444893 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:43:32 crc kubenswrapper[4741]: I0929 21:43:32.463295 4741 generic.go:334] "Generic (PLEG): container finished" podID="719f915f-590f-4363-a719-686c7dc3f9fb" containerID="1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4" exitCode=0 Sep 29 21:43:32 crc kubenswrapper[4741]: I0929 21:43:32.463506 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvp42" event={"ID":"719f915f-590f-4363-a719-686c7dc3f9fb","Type":"ContainerDied","Data":"1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4"} Sep 29 21:43:32 crc kubenswrapper[4741]: I0929 21:43:32.469204 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85nhj" event={"ID":"32dfad7c-e541-4a01-a3d0-2520ffd05128","Type":"ContainerStarted","Data":"bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044"} Sep 29 21:43:33 crc kubenswrapper[4741]: I0929 21:43:33.490883 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvp42" event={"ID":"719f915f-590f-4363-a719-686c7dc3f9fb","Type":"ContainerStarted","Data":"9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8"} Sep 29 21:43:34 crc kubenswrapper[4741]: I0929 21:43:34.499628 4741 generic.go:334] "Generic (PLEG): container finished" podID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerID="bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044" exitCode=0 Sep 29 21:43:34 crc kubenswrapper[4741]: I0929 21:43:34.499717 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85nhj" event={"ID":"32dfad7c-e541-4a01-a3d0-2520ffd05128","Type":"ContainerDied","Data":"bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044"} Sep 29 21:43:36 crc kubenswrapper[4741]: I0929 21:43:36.525455 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85nhj" event={"ID":"32dfad7c-e541-4a01-a3d0-2520ffd05128","Type":"ContainerStarted","Data":"48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f"} Sep 29 21:43:36 crc kubenswrapper[4741]: I0929 21:43:36.527892 4741 generic.go:334] "Generic (PLEG): container finished" podID="719f915f-590f-4363-a719-686c7dc3f9fb" containerID="9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8" exitCode=0 Sep 29 21:43:36 crc kubenswrapper[4741]: I0929 21:43:36.527932 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvp42" event={"ID":"719f915f-590f-4363-a719-686c7dc3f9fb","Type":"ContainerDied","Data":"9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8"} Sep 29 21:43:36 crc kubenswrapper[4741]: I0929 21:43:36.552863 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-85nhj" podStartSLOduration=2.662339138 podStartE2EDuration="6.552825004s" podCreationTimestamp="2025-09-29 21:43:30 +0000 UTC" firstStartedPulling="2025-09-29 21:43:31.444648107 +0000 UTC m=+9253.092437439" lastFinishedPulling="2025-09-29 21:43:35.335133963 +0000 UTC m=+9256.982923305" observedRunningTime="2025-09-29 21:43:36.543702687 +0000 UTC m=+9258.191492019" watchObservedRunningTime="2025-09-29 21:43:36.552825004 +0000 UTC m=+9258.200614336" Sep 29 21:43:37 crc kubenswrapper[4741]: I0929 21:43:37.538379 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvp42" event={"ID":"719f915f-590f-4363-a719-686c7dc3f9fb","Type":"ContainerStarted","Data":"1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c"} Sep 29 21:43:40 crc kubenswrapper[4741]: I0929 21:43:40.444800 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:40 crc kubenswrapper[4741]: I0929 21:43:40.445297 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:40 crc kubenswrapper[4741]: I0929 21:43:40.645418 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:40 crc kubenswrapper[4741]: I0929 21:43:40.645503 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:40 crc kubenswrapper[4741]: I0929 21:43:40.693886 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:40 crc kubenswrapper[4741]: I0929 21:43:40.719710 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dvp42" podStartSLOduration=5.999109849 podStartE2EDuration="10.719686589s" podCreationTimestamp="2025-09-29 21:43:30 +0000 UTC" firstStartedPulling="2025-09-29 21:43:32.465332448 +0000 UTC m=+9254.113121780" lastFinishedPulling="2025-09-29 21:43:37.185909188 +0000 UTC m=+9258.833698520" observedRunningTime="2025-09-29 21:43:37.562794578 +0000 UTC m=+9259.210583910" watchObservedRunningTime="2025-09-29 21:43:40.719686589 +0000 UTC m=+9262.367475921" Sep 29 21:43:41 crc kubenswrapper[4741]: I0929 21:43:41.496867 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-85nhj" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="registry-server" probeResult="failure" output=< Sep 29 21:43:41 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 21:43:41 crc kubenswrapper[4741]: > Sep 29 21:43:42 crc kubenswrapper[4741]: I0929 21:43:42.657264 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:42 crc kubenswrapper[4741]: I0929 21:43:42.909514 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dvp42"] Sep 29 21:43:44 crc kubenswrapper[4741]: I0929 21:43:44.611274 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dvp42" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="registry-server" containerID="cri-o://1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c" gracePeriod=2 Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.177489 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.294828 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-utilities\") pod \"719f915f-590f-4363-a719-686c7dc3f9fb\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.294980 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4js5t\" (UniqueName: \"kubernetes.io/projected/719f915f-590f-4363-a719-686c7dc3f9fb-kube-api-access-4js5t\") pod \"719f915f-590f-4363-a719-686c7dc3f9fb\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.295158 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-catalog-content\") pod \"719f915f-590f-4363-a719-686c7dc3f9fb\" (UID: \"719f915f-590f-4363-a719-686c7dc3f9fb\") " Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.296873 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-utilities" (OuterVolumeSpecName: "utilities") pod "719f915f-590f-4363-a719-686c7dc3f9fb" (UID: "719f915f-590f-4363-a719-686c7dc3f9fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.304436 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/719f915f-590f-4363-a719-686c7dc3f9fb-kube-api-access-4js5t" (OuterVolumeSpecName: "kube-api-access-4js5t") pod "719f915f-590f-4363-a719-686c7dc3f9fb" (UID: "719f915f-590f-4363-a719-686c7dc3f9fb"). InnerVolumeSpecName "kube-api-access-4js5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.341269 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "719f915f-590f-4363-a719-686c7dc3f9fb" (UID: "719f915f-590f-4363-a719-686c7dc3f9fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.398554 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.398592 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4js5t\" (UniqueName: \"kubernetes.io/projected/719f915f-590f-4363-a719-686c7dc3f9fb-kube-api-access-4js5t\") on node \"crc\" DevicePath \"\"" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.398604 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/719f915f-590f-4363-a719-686c7dc3f9fb-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.625223 4741 generic.go:334] "Generic (PLEG): container finished" podID="719f915f-590f-4363-a719-686c7dc3f9fb" containerID="1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c" exitCode=0 Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.625292 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvp42" event={"ID":"719f915f-590f-4363-a719-686c7dc3f9fb","Type":"ContainerDied","Data":"1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c"} Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.625783 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvp42" event={"ID":"719f915f-590f-4363-a719-686c7dc3f9fb","Type":"ContainerDied","Data":"e2387587ee1c51a4d5ee9d2018323601f9f3df2af0275c393d7f8044c1771e53"} Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.625817 4741 scope.go:117] "RemoveContainer" containerID="1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.625336 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvp42" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.665369 4741 scope.go:117] "RemoveContainer" containerID="9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.672978 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dvp42"] Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.688686 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dvp42"] Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.703193 4741 scope.go:117] "RemoveContainer" containerID="1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.742240 4741 scope.go:117] "RemoveContainer" containerID="1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c" Sep 29 21:43:45 crc kubenswrapper[4741]: E0929 21:43:45.742793 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c\": container with ID starting with 1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c not found: ID does not exist" containerID="1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.742839 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c"} err="failed to get container status \"1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c\": rpc error: code = NotFound desc = could not find container \"1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c\": container with ID starting with 1116d5d864855553705dd424d6ad3f508509542a002c5df39ddbd9df0ddc5c5c not found: ID does not exist" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.742871 4741 scope.go:117] "RemoveContainer" containerID="9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8" Sep 29 21:43:45 crc kubenswrapper[4741]: E0929 21:43:45.743168 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8\": container with ID starting with 9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8 not found: ID does not exist" containerID="9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.743188 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8"} err="failed to get container status \"9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8\": rpc error: code = NotFound desc = could not find container \"9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8\": container with ID starting with 9271ee7d24427c798405ac80faf62deb12508520bc901b3e5e1fbe4307a967e8 not found: ID does not exist" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.743205 4741 scope.go:117] "RemoveContainer" containerID="1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4" Sep 29 21:43:45 crc kubenswrapper[4741]: E0929 21:43:45.743450 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4\": container with ID starting with 1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4 not found: ID does not exist" containerID="1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4" Sep 29 21:43:45 crc kubenswrapper[4741]: I0929 21:43:45.743475 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4"} err="failed to get container status \"1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4\": rpc error: code = NotFound desc = could not find container \"1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4\": container with ID starting with 1bb9526833c8a2a9ea028a7963607d97fd284825ddcca339c07003bddfeeeed4 not found: ID does not exist" Sep 29 21:43:47 crc kubenswrapper[4741]: I0929 21:43:47.097438 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" path="/var/lib/kubelet/pods/719f915f-590f-4363-a719-686c7dc3f9fb/volumes" Sep 29 21:43:50 crc kubenswrapper[4741]: I0929 21:43:50.493577 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:50 crc kubenswrapper[4741]: I0929 21:43:50.547001 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:50 crc kubenswrapper[4741]: I0929 21:43:50.732156 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85nhj"] Sep 29 21:43:51 crc kubenswrapper[4741]: I0929 21:43:51.683664 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-85nhj" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="registry-server" containerID="cri-o://48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f" gracePeriod=2 Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.265310 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.374785 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq72t\" (UniqueName: \"kubernetes.io/projected/32dfad7c-e541-4a01-a3d0-2520ffd05128-kube-api-access-cq72t\") pod \"32dfad7c-e541-4a01-a3d0-2520ffd05128\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.374889 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-utilities\") pod \"32dfad7c-e541-4a01-a3d0-2520ffd05128\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.374957 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-catalog-content\") pod \"32dfad7c-e541-4a01-a3d0-2520ffd05128\" (UID: \"32dfad7c-e541-4a01-a3d0-2520ffd05128\") " Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.376356 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-utilities" (OuterVolumeSpecName: "utilities") pod "32dfad7c-e541-4a01-a3d0-2520ffd05128" (UID: "32dfad7c-e541-4a01-a3d0-2520ffd05128"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.380616 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32dfad7c-e541-4a01-a3d0-2520ffd05128-kube-api-access-cq72t" (OuterVolumeSpecName: "kube-api-access-cq72t") pod "32dfad7c-e541-4a01-a3d0-2520ffd05128" (UID: "32dfad7c-e541-4a01-a3d0-2520ffd05128"). InnerVolumeSpecName "kube-api-access-cq72t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.464240 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32dfad7c-e541-4a01-a3d0-2520ffd05128" (UID: "32dfad7c-e541-4a01-a3d0-2520ffd05128"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.477686 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq72t\" (UniqueName: \"kubernetes.io/projected/32dfad7c-e541-4a01-a3d0-2520ffd05128-kube-api-access-cq72t\") on node \"crc\" DevicePath \"\"" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.477913 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.477982 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dfad7c-e541-4a01-a3d0-2520ffd05128-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.699218 4741 generic.go:334] "Generic (PLEG): container finished" podID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerID="48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f" exitCode=0 Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.699682 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85nhj" event={"ID":"32dfad7c-e541-4a01-a3d0-2520ffd05128","Type":"ContainerDied","Data":"48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f"} Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.699727 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85nhj" event={"ID":"32dfad7c-e541-4a01-a3d0-2520ffd05128","Type":"ContainerDied","Data":"df0554b1429657f1e472d26049a94557cfdb72780c03e78785875ea25fdfbd05"} Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.699756 4741 scope.go:117] "RemoveContainer" containerID="48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.699989 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85nhj" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.723251 4741 scope.go:117] "RemoveContainer" containerID="bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.750779 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85nhj"] Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.762414 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-85nhj"] Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.765111 4741 scope.go:117] "RemoveContainer" containerID="58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.796774 4741 scope.go:117] "RemoveContainer" containerID="48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f" Sep 29 21:43:52 crc kubenswrapper[4741]: E0929 21:43:52.797155 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f\": container with ID starting with 48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f not found: ID does not exist" containerID="48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.797188 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f"} err="failed to get container status \"48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f\": rpc error: code = NotFound desc = could not find container \"48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f\": container with ID starting with 48c9d778cbb95f5576856b44f14d2898b674dfc3f45263986264910a80bad63f not found: ID does not exist" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.797211 4741 scope.go:117] "RemoveContainer" containerID="bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044" Sep 29 21:43:52 crc kubenswrapper[4741]: E0929 21:43:52.797420 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044\": container with ID starting with bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044 not found: ID does not exist" containerID="bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.797444 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044"} err="failed to get container status \"bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044\": rpc error: code = NotFound desc = could not find container \"bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044\": container with ID starting with bdffa82e88cf8d86b9419627779b95ea96994dfee93688d6aaff1cf90a053044 not found: ID does not exist" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.797460 4741 scope.go:117] "RemoveContainer" containerID="58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e" Sep 29 21:43:52 crc kubenswrapper[4741]: E0929 21:43:52.797823 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e\": container with ID starting with 58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e not found: ID does not exist" containerID="58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e" Sep 29 21:43:52 crc kubenswrapper[4741]: I0929 21:43:52.797882 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e"} err="failed to get container status \"58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e\": rpc error: code = NotFound desc = could not find container \"58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e\": container with ID starting with 58de05aac9d8fa54976fed3da415947de91d4d21c6a87251ce0824ba33f5d03e not found: ID does not exist" Sep 29 21:43:53 crc kubenswrapper[4741]: I0929 21:43:53.097417 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" path="/var/lib/kubelet/pods/32dfad7c-e541-4a01-a3d0-2520ffd05128/volumes" Sep 29 21:44:01 crc kubenswrapper[4741]: I0929 21:44:01.738893 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:44:01 crc kubenswrapper[4741]: I0929 21:44:01.739435 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:44:31 crc kubenswrapper[4741]: I0929 21:44:31.738771 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:44:31 crc kubenswrapper[4741]: I0929 21:44:31.739164 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.159503 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319705-77994"] Sep 29 21:45:00 crc kubenswrapper[4741]: E0929 21:45:00.160547 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="extract-content" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160560 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="extract-content" Sep 29 21:45:00 crc kubenswrapper[4741]: E0929 21:45:00.160573 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="extract-utilities" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160579 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="extract-utilities" Sep 29 21:45:00 crc kubenswrapper[4741]: E0929 21:45:00.160588 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="extract-utilities" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160594 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="extract-utilities" Sep 29 21:45:00 crc kubenswrapper[4741]: E0929 21:45:00.160603 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="registry-server" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160609 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="registry-server" Sep 29 21:45:00 crc kubenswrapper[4741]: E0929 21:45:00.160623 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="extract-content" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160628 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="extract-content" Sep 29 21:45:00 crc kubenswrapper[4741]: E0929 21:45:00.160641 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="registry-server" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160647 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="registry-server" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160839 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="32dfad7c-e541-4a01-a3d0-2520ffd05128" containerName="registry-server" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.160865 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="719f915f-590f-4363-a719-686c7dc3f9fb" containerName="registry-server" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.161692 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.168363 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.169479 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.172246 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319705-77994"] Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.245706 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b49dced-510f-4b2d-b622-b51d503a5476-secret-volume\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.245810 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b49dced-510f-4b2d-b622-b51d503a5476-config-volume\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.245832 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gb8g\" (UniqueName: \"kubernetes.io/projected/1b49dced-510f-4b2d-b622-b51d503a5476-kube-api-access-4gb8g\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.347536 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b49dced-510f-4b2d-b622-b51d503a5476-secret-volume\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.347663 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gb8g\" (UniqueName: \"kubernetes.io/projected/1b49dced-510f-4b2d-b622-b51d503a5476-kube-api-access-4gb8g\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.347684 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b49dced-510f-4b2d-b622-b51d503a5476-config-volume\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.348833 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b49dced-510f-4b2d-b622-b51d503a5476-config-volume\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.356513 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b49dced-510f-4b2d-b622-b51d503a5476-secret-volume\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.367125 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gb8g\" (UniqueName: \"kubernetes.io/projected/1b49dced-510f-4b2d-b622-b51d503a5476-kube-api-access-4gb8g\") pod \"collect-profiles-29319705-77994\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.487491 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:00 crc kubenswrapper[4741]: I0929 21:45:00.939682 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319705-77994"] Sep 29 21:45:01 crc kubenswrapper[4741]: I0929 21:45:01.404079 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" event={"ID":"1b49dced-510f-4b2d-b622-b51d503a5476","Type":"ContainerStarted","Data":"248ca02ee6080d7af9c685ec87bb26303ceb5210ba92337faa360903dda41490"} Sep 29 21:45:01 crc kubenswrapper[4741]: I0929 21:45:01.738688 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:45:01 crc kubenswrapper[4741]: I0929 21:45:01.738743 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:45:01 crc kubenswrapper[4741]: I0929 21:45:01.738779 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:45:01 crc kubenswrapper[4741]: I0929 21:45:01.739436 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a12d2eb3bbc17d7730f91f2eaa48266afea0ad8837e02e3139eea4a70466b1a"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:45:01 crc kubenswrapper[4741]: I0929 21:45:01.739491 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://3a12d2eb3bbc17d7730f91f2eaa48266afea0ad8837e02e3139eea4a70466b1a" gracePeriod=600 Sep 29 21:45:02 crc kubenswrapper[4741]: I0929 21:45:02.414867 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="3a12d2eb3bbc17d7730f91f2eaa48266afea0ad8837e02e3139eea4a70466b1a" exitCode=0 Sep 29 21:45:02 crc kubenswrapper[4741]: I0929 21:45:02.414971 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"3a12d2eb3bbc17d7730f91f2eaa48266afea0ad8837e02e3139eea4a70466b1a"} Sep 29 21:45:02 crc kubenswrapper[4741]: I0929 21:45:02.415243 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90"} Sep 29 21:45:02 crc kubenswrapper[4741]: I0929 21:45:02.415267 4741 scope.go:117] "RemoveContainer" containerID="0ef6f94e8246bb578d8c505504dd9fca0ff46531e6f3f06d423a82c6baa45fd2" Sep 29 21:45:02 crc kubenswrapper[4741]: I0929 21:45:02.417428 4741 generic.go:334] "Generic (PLEG): container finished" podID="1b49dced-510f-4b2d-b622-b51d503a5476" containerID="3c8895f01194995fa53fd2ccae16c24bf147951a55fe4759d41ea330b62ac939" exitCode=0 Sep 29 21:45:02 crc kubenswrapper[4741]: I0929 21:45:02.417459 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" event={"ID":"1b49dced-510f-4b2d-b622-b51d503a5476","Type":"ContainerDied","Data":"3c8895f01194995fa53fd2ccae16c24bf147951a55fe4759d41ea330b62ac939"} Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.771166 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.819312 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b49dced-510f-4b2d-b622-b51d503a5476-secret-volume\") pod \"1b49dced-510f-4b2d-b622-b51d503a5476\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.819474 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b49dced-510f-4b2d-b622-b51d503a5476-config-volume\") pod \"1b49dced-510f-4b2d-b622-b51d503a5476\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.819551 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gb8g\" (UniqueName: \"kubernetes.io/projected/1b49dced-510f-4b2d-b622-b51d503a5476-kube-api-access-4gb8g\") pod \"1b49dced-510f-4b2d-b622-b51d503a5476\" (UID: \"1b49dced-510f-4b2d-b622-b51d503a5476\") " Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.820590 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b49dced-510f-4b2d-b622-b51d503a5476-config-volume" (OuterVolumeSpecName: "config-volume") pod "1b49dced-510f-4b2d-b622-b51d503a5476" (UID: "1b49dced-510f-4b2d-b622-b51d503a5476"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.820834 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b49dced-510f-4b2d-b622-b51d503a5476-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.826691 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b49dced-510f-4b2d-b622-b51d503a5476-kube-api-access-4gb8g" (OuterVolumeSpecName: "kube-api-access-4gb8g") pod "1b49dced-510f-4b2d-b622-b51d503a5476" (UID: "1b49dced-510f-4b2d-b622-b51d503a5476"). InnerVolumeSpecName "kube-api-access-4gb8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.827544 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49dced-510f-4b2d-b622-b51d503a5476-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1b49dced-510f-4b2d-b622-b51d503a5476" (UID: "1b49dced-510f-4b2d-b622-b51d503a5476"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.922267 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b49dced-510f-4b2d-b622-b51d503a5476-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 21:45:03 crc kubenswrapper[4741]: I0929 21:45:03.922553 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gb8g\" (UniqueName: \"kubernetes.io/projected/1b49dced-510f-4b2d-b622-b51d503a5476-kube-api-access-4gb8g\") on node \"crc\" DevicePath \"\"" Sep 29 21:45:04 crc kubenswrapper[4741]: I0929 21:45:04.441285 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" event={"ID":"1b49dced-510f-4b2d-b622-b51d503a5476","Type":"ContainerDied","Data":"248ca02ee6080d7af9c685ec87bb26303ceb5210ba92337faa360903dda41490"} Sep 29 21:45:04 crc kubenswrapper[4741]: I0929 21:45:04.441338 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248ca02ee6080d7af9c685ec87bb26303ceb5210ba92337faa360903dda41490" Sep 29 21:45:04 crc kubenswrapper[4741]: I0929 21:45:04.441373 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319705-77994" Sep 29 21:45:04 crc kubenswrapper[4741]: I0929 21:45:04.850248 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz"] Sep 29 21:45:04 crc kubenswrapper[4741]: I0929 21:45:04.862025 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319660-nl4kz"] Sep 29 21:45:05 crc kubenswrapper[4741]: I0929 21:45:05.099811 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="318cac1a-c05b-4e3c-a379-cbcbbf8cd071" path="/var/lib/kubelet/pods/318cac1a-c05b-4e3c-a379-cbcbbf8cd071/volumes" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.828584 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z6hpq"] Sep 29 21:45:19 crc kubenswrapper[4741]: E0929 21:45:19.830731 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b49dced-510f-4b2d-b622-b51d503a5476" containerName="collect-profiles" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.830752 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b49dced-510f-4b2d-b622-b51d503a5476" containerName="collect-profiles" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.830988 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b49dced-510f-4b2d-b622-b51d503a5476" containerName="collect-profiles" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.832962 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.843410 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6hpq"] Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.883822 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d6cr\" (UniqueName: \"kubernetes.io/projected/f770b139-f86c-474a-942d-d8658a928c3f-kube-api-access-8d6cr\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.884084 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-utilities\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.884239 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-catalog-content\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.985810 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-catalog-content\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.986473 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-catalog-content\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.986636 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d6cr\" (UniqueName: \"kubernetes.io/projected/f770b139-f86c-474a-942d-d8658a928c3f-kube-api-access-8d6cr\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.986768 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-utilities\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:19 crc kubenswrapper[4741]: I0929 21:45:19.987150 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-utilities\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:20 crc kubenswrapper[4741]: I0929 21:45:20.007901 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d6cr\" (UniqueName: \"kubernetes.io/projected/f770b139-f86c-474a-942d-d8658a928c3f-kube-api-access-8d6cr\") pod \"community-operators-z6hpq\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:20 crc kubenswrapper[4741]: I0929 21:45:20.161433 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:20 crc kubenswrapper[4741]: I0929 21:45:20.697093 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6hpq"] Sep 29 21:45:21 crc kubenswrapper[4741]: I0929 21:45:21.639935 4741 generic.go:334] "Generic (PLEG): container finished" podID="f770b139-f86c-474a-942d-d8658a928c3f" containerID="a3798b7539bb3657889f53b3e495469e3786265e3e3ed8325a82ca2ca1ac24bc" exitCode=0 Sep 29 21:45:21 crc kubenswrapper[4741]: I0929 21:45:21.640156 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6hpq" event={"ID":"f770b139-f86c-474a-942d-d8658a928c3f","Type":"ContainerDied","Data":"a3798b7539bb3657889f53b3e495469e3786265e3e3ed8325a82ca2ca1ac24bc"} Sep 29 21:45:21 crc kubenswrapper[4741]: I0929 21:45:21.640494 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6hpq" event={"ID":"f770b139-f86c-474a-942d-d8658a928c3f","Type":"ContainerStarted","Data":"17fe8aac5ff63196fda50e0713cf388ed62388371c34bd911b7b735777235bdc"} Sep 29 21:45:22 crc kubenswrapper[4741]: I0929 21:45:22.654045 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6hpq" event={"ID":"f770b139-f86c-474a-942d-d8658a928c3f","Type":"ContainerStarted","Data":"0797f35f0c7205bf8d2a406f493148e0d58f4a36524e9ac6cefab38414a382f0"} Sep 29 21:45:23 crc kubenswrapper[4741]: I0929 21:45:23.665669 4741 generic.go:334] "Generic (PLEG): container finished" podID="f770b139-f86c-474a-942d-d8658a928c3f" containerID="0797f35f0c7205bf8d2a406f493148e0d58f4a36524e9ac6cefab38414a382f0" exitCode=0 Sep 29 21:45:23 crc kubenswrapper[4741]: I0929 21:45:23.665789 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6hpq" event={"ID":"f770b139-f86c-474a-942d-d8658a928c3f","Type":"ContainerDied","Data":"0797f35f0c7205bf8d2a406f493148e0d58f4a36524e9ac6cefab38414a382f0"} Sep 29 21:45:24 crc kubenswrapper[4741]: I0929 21:45:24.677780 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6hpq" event={"ID":"f770b139-f86c-474a-942d-d8658a928c3f","Type":"ContainerStarted","Data":"0e2a2117ea7e6220a78a106271775f47044acb5521513b16dbd10785f7635629"} Sep 29 21:45:24 crc kubenswrapper[4741]: I0929 21:45:24.701295 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z6hpq" podStartSLOduration=3.104319348 podStartE2EDuration="5.701279097s" podCreationTimestamp="2025-09-29 21:45:19 +0000 UTC" firstStartedPulling="2025-09-29 21:45:21.644729694 +0000 UTC m=+9363.292519066" lastFinishedPulling="2025-09-29 21:45:24.241689483 +0000 UTC m=+9365.889478815" observedRunningTime="2025-09-29 21:45:24.694005317 +0000 UTC m=+9366.341794649" watchObservedRunningTime="2025-09-29 21:45:24.701279097 +0000 UTC m=+9366.349068429" Sep 29 21:45:30 crc kubenswrapper[4741]: I0929 21:45:30.161750 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:30 crc kubenswrapper[4741]: I0929 21:45:30.164362 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:30 crc kubenswrapper[4741]: I0929 21:45:30.216452 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:30 crc kubenswrapper[4741]: I0929 21:45:30.789526 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:33 crc kubenswrapper[4741]: I0929 21:45:33.032020 4741 scope.go:117] "RemoveContainer" containerID="c7a38977ae13f188cb856e63905e8deb59e17ee82118871d26e8676cd6658eca" Sep 29 21:45:34 crc kubenswrapper[4741]: I0929 21:45:34.019885 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z6hpq"] Sep 29 21:45:34 crc kubenswrapper[4741]: I0929 21:45:34.020081 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z6hpq" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="registry-server" containerID="cri-o://0e2a2117ea7e6220a78a106271775f47044acb5521513b16dbd10785f7635629" gracePeriod=2 Sep 29 21:45:34 crc kubenswrapper[4741]: I0929 21:45:34.777105 4741 generic.go:334] "Generic (PLEG): container finished" podID="f770b139-f86c-474a-942d-d8658a928c3f" containerID="0e2a2117ea7e6220a78a106271775f47044acb5521513b16dbd10785f7635629" exitCode=0 Sep 29 21:45:34 crc kubenswrapper[4741]: I0929 21:45:34.777187 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6hpq" event={"ID":"f770b139-f86c-474a-942d-d8658a928c3f","Type":"ContainerDied","Data":"0e2a2117ea7e6220a78a106271775f47044acb5521513b16dbd10785f7635629"} Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.191471 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.220349 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-catalog-content\") pod \"f770b139-f86c-474a-942d-d8658a928c3f\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.220483 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-utilities\") pod \"f770b139-f86c-474a-942d-d8658a928c3f\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.220767 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d6cr\" (UniqueName: \"kubernetes.io/projected/f770b139-f86c-474a-942d-d8658a928c3f-kube-api-access-8d6cr\") pod \"f770b139-f86c-474a-942d-d8658a928c3f\" (UID: \"f770b139-f86c-474a-942d-d8658a928c3f\") " Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.221413 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-utilities" (OuterVolumeSpecName: "utilities") pod "f770b139-f86c-474a-942d-d8658a928c3f" (UID: "f770b139-f86c-474a-942d-d8658a928c3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.227309 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f770b139-f86c-474a-942d-d8658a928c3f-kube-api-access-8d6cr" (OuterVolumeSpecName: "kube-api-access-8d6cr") pod "f770b139-f86c-474a-942d-d8658a928c3f" (UID: "f770b139-f86c-474a-942d-d8658a928c3f"). InnerVolumeSpecName "kube-api-access-8d6cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.272309 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f770b139-f86c-474a-942d-d8658a928c3f" (UID: "f770b139-f86c-474a-942d-d8658a928c3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.322857 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d6cr\" (UniqueName: \"kubernetes.io/projected/f770b139-f86c-474a-942d-d8658a928c3f-kube-api-access-8d6cr\") on node \"crc\" DevicePath \"\"" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.322908 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.322932 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f770b139-f86c-474a-942d-d8658a928c3f-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.788976 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6hpq" event={"ID":"f770b139-f86c-474a-942d-d8658a928c3f","Type":"ContainerDied","Data":"17fe8aac5ff63196fda50e0713cf388ed62388371c34bd911b7b735777235bdc"} Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.789037 4741 scope.go:117] "RemoveContainer" containerID="0e2a2117ea7e6220a78a106271775f47044acb5521513b16dbd10785f7635629" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.789094 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6hpq" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.820321 4741 scope.go:117] "RemoveContainer" containerID="0797f35f0c7205bf8d2a406f493148e0d58f4a36524e9ac6cefab38414a382f0" Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.823790 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z6hpq"] Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.832559 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z6hpq"] Sep 29 21:45:35 crc kubenswrapper[4741]: I0929 21:45:35.852497 4741 scope.go:117] "RemoveContainer" containerID="a3798b7539bb3657889f53b3e495469e3786265e3e3ed8325a82ca2ca1ac24bc" Sep 29 21:45:37 crc kubenswrapper[4741]: I0929 21:45:37.097601 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f770b139-f86c-474a-942d-d8658a928c3f" path="/var/lib/kubelet/pods/f770b139-f86c-474a-942d-d8658a928c3f/volumes" Sep 29 21:47:31 crc kubenswrapper[4741]: I0929 21:47:31.738408 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:47:31 crc kubenswrapper[4741]: I0929 21:47:31.738923 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:48:01 crc kubenswrapper[4741]: I0929 21:48:01.739435 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:48:01 crc kubenswrapper[4741]: I0929 21:48:01.739989 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:48:31 crc kubenswrapper[4741]: I0929 21:48:31.739433 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:48:31 crc kubenswrapper[4741]: I0929 21:48:31.740196 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:48:31 crc kubenswrapper[4741]: I0929 21:48:31.740257 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:48:31 crc kubenswrapper[4741]: I0929 21:48:31.741522 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:48:31 crc kubenswrapper[4741]: I0929 21:48:31.741624 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" gracePeriod=600 Sep 29 21:48:31 crc kubenswrapper[4741]: E0929 21:48:31.867375 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:48:32 crc kubenswrapper[4741]: I0929 21:48:32.542570 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" exitCode=0 Sep 29 21:48:32 crc kubenswrapper[4741]: I0929 21:48:32.542627 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90"} Sep 29 21:48:32 crc kubenswrapper[4741]: I0929 21:48:32.542660 4741 scope.go:117] "RemoveContainer" containerID="3a12d2eb3bbc17d7730f91f2eaa48266afea0ad8837e02e3139eea4a70466b1a" Sep 29 21:48:32 crc kubenswrapper[4741]: I0929 21:48:32.543397 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:48:32 crc kubenswrapper[4741]: E0929 21:48:32.543713 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:48:45 crc kubenswrapper[4741]: I0929 21:48:45.087322 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:48:45 crc kubenswrapper[4741]: E0929 21:48:45.088193 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:48:57 crc kubenswrapper[4741]: I0929 21:48:57.086452 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:48:57 crc kubenswrapper[4741]: E0929 21:48:57.088410 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:49:09 crc kubenswrapper[4741]: I0929 21:49:09.117460 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:49:09 crc kubenswrapper[4741]: E0929 21:49:09.118325 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:49:22 crc kubenswrapper[4741]: I0929 21:49:22.085958 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:49:22 crc kubenswrapper[4741]: E0929 21:49:22.086890 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:49:35 crc kubenswrapper[4741]: I0929 21:49:35.087315 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:49:35 crc kubenswrapper[4741]: E0929 21:49:35.088098 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:49:50 crc kubenswrapper[4741]: I0929 21:49:50.086270 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:49:50 crc kubenswrapper[4741]: E0929 21:49:50.086995 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:50:05 crc kubenswrapper[4741]: I0929 21:50:05.086906 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:50:05 crc kubenswrapper[4741]: E0929 21:50:05.087969 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:50:19 crc kubenswrapper[4741]: I0929 21:50:19.094684 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:50:19 crc kubenswrapper[4741]: E0929 21:50:19.095425 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:50:30 crc kubenswrapper[4741]: I0929 21:50:30.086809 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:50:30 crc kubenswrapper[4741]: E0929 21:50:30.087790 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:50:41 crc kubenswrapper[4741]: I0929 21:50:41.086857 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:50:41 crc kubenswrapper[4741]: E0929 21:50:41.087878 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:50:56 crc kubenswrapper[4741]: I0929 21:50:56.086635 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:50:56 crc kubenswrapper[4741]: E0929 21:50:56.087484 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:51:07 crc kubenswrapper[4741]: I0929 21:51:07.085944 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:51:07 crc kubenswrapper[4741]: E0929 21:51:07.086745 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:51:12 crc kubenswrapper[4741]: I0929 21:51:12.303202 4741 generic.go:334] "Generic (PLEG): container finished" podID="b6c8b549-a934-4c9d-919c-b5dee30d4385" containerID="b927dc65d5526a3b4c8678a09a011b8e934c4458c0c6f854ce6018f56e465ac8" exitCode=0 Sep 29 21:51:12 crc kubenswrapper[4741]: I0929 21:51:12.303279 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" event={"ID":"b6c8b549-a934-4c9d-919c-b5dee30d4385","Type":"ContainerDied","Data":"b927dc65d5526a3b4c8678a09a011b8e934c4458c0c6f854ce6018f56e465ac8"} Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.786417 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.971946 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxzbc\" (UniqueName: \"kubernetes.io/projected/b6c8b549-a934-4c9d-919c-b5dee30d4385-kube-api-access-mxzbc\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972050 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ssh-key\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972117 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-1\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972161 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-inventory\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972195 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-1\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972237 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-combined-ca-bundle\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972313 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-0\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972373 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-0\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972422 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ceph\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972478 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-1\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.972532 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-0\") pod \"b6c8b549-a934-4c9d-919c-b5dee30d4385\" (UID: \"b6c8b549-a934-4c9d-919c-b5dee30d4385\") " Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.977765 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ceph" (OuterVolumeSpecName: "ceph") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.979370 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c8b549-a934-4c9d-919c-b5dee30d4385-kube-api-access-mxzbc" (OuterVolumeSpecName: "kube-api-access-mxzbc") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "kube-api-access-mxzbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:51:13 crc kubenswrapper[4741]: I0929 21:51:13.982546 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.002079 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.005464 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.014960 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.017257 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.020677 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.027464 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.030611 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.043204 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-inventory" (OuterVolumeSpecName: "inventory") pod "b6c8b549-a934-4c9d-919c-b5dee30d4385" (UID: "b6c8b549-a934-4c9d-919c-b5dee30d4385"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076250 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076316 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076328 4741 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ceph\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076338 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076414 4741 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076454 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxzbc\" (UniqueName: \"kubernetes.io/projected/b6c8b549-a934-4c9d-919c-b5dee30d4385-kube-api-access-mxzbc\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076464 4741 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076474 4741 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076483 4741 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076492 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.076500 4741 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c8b549-a934-4c9d-919c-b5dee30d4385-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.325124 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" event={"ID":"b6c8b549-a934-4c9d-919c-b5dee30d4385","Type":"ContainerDied","Data":"b9aacf87b00118ac4a98a8a8f4a51ba5c84a4b184ea1ff775c35e9730ac84d47"} Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.325466 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9aacf87b00118ac4a98a8a8f4a51ba5c84a4b184ea1ff775c35e9730ac84d47" Sep 29 21:51:14 crc kubenswrapper[4741]: I0929 21:51:14.325212 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8" Sep 29 21:51:22 crc kubenswrapper[4741]: I0929 21:51:22.087372 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:51:22 crc kubenswrapper[4741]: E0929 21:51:22.088095 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:51:34 crc kubenswrapper[4741]: I0929 21:51:34.085745 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:51:34 crc kubenswrapper[4741]: E0929 21:51:34.086570 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:51:48 crc kubenswrapper[4741]: I0929 21:51:48.086730 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:51:48 crc kubenswrapper[4741]: E0929 21:51:48.087583 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:51:59 crc kubenswrapper[4741]: I0929 21:51:59.094115 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:51:59 crc kubenswrapper[4741]: E0929 21:51:59.094876 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:52:13 crc kubenswrapper[4741]: I0929 21:52:13.086729 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:52:13 crc kubenswrapper[4741]: E0929 21:52:13.087547 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:52:26 crc kubenswrapper[4741]: I0929 21:52:26.086543 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:52:26 crc kubenswrapper[4741]: E0929 21:52:26.087292 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:52:39 crc kubenswrapper[4741]: I0929 21:52:39.092596 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:52:39 crc kubenswrapper[4741]: E0929 21:52:39.093201 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:52:50 crc kubenswrapper[4741]: I0929 21:52:50.086018 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:52:50 crc kubenswrapper[4741]: E0929 21:52:50.087116 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:53:04 crc kubenswrapper[4741]: I0929 21:53:04.086942 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:53:04 crc kubenswrapper[4741]: E0929 21:53:04.088786 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:53:15 crc kubenswrapper[4741]: I0929 21:53:15.086612 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:53:15 crc kubenswrapper[4741]: E0929 21:53:15.087256 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.252568 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-chpfj"] Sep 29 21:53:24 crc kubenswrapper[4741]: E0929 21:53:24.253568 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c8b549-a934-4c9d-919c-b5dee30d4385" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.253583 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c8b549-a934-4c9d-919c-b5dee30d4385" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Sep 29 21:53:24 crc kubenswrapper[4741]: E0929 21:53:24.253607 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="extract-utilities" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.253614 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="extract-utilities" Sep 29 21:53:24 crc kubenswrapper[4741]: E0929 21:53:24.253633 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="registry-server" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.253639 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="registry-server" Sep 29 21:53:24 crc kubenswrapper[4741]: E0929 21:53:24.253664 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="extract-content" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.253671 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="extract-content" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.253888 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c8b549-a934-4c9d-919c-b5dee30d4385" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.253898 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="f770b139-f86c-474a-942d-d8658a928c3f" containerName="registry-server" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.255544 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.261500 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-chpfj"] Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.426211 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-utilities\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.426717 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpkzm\" (UniqueName: \"kubernetes.io/projected/fdade596-ab6b-41ae-8c25-2f848ace7f87-kube-api-access-lpkzm\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.426999 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-catalog-content\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.529841 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-utilities\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.529897 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpkzm\" (UniqueName: \"kubernetes.io/projected/fdade596-ab6b-41ae-8c25-2f848ace7f87-kube-api-access-lpkzm\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.529952 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-catalog-content\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.530669 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-catalog-content\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.531354 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-utilities\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.551067 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpkzm\" (UniqueName: \"kubernetes.io/projected/fdade596-ab6b-41ae-8c25-2f848ace7f87-kube-api-access-lpkzm\") pod \"redhat-marketplace-chpfj\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:24 crc kubenswrapper[4741]: I0929 21:53:24.590949 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:25 crc kubenswrapper[4741]: I0929 21:53:25.053045 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-chpfj"] Sep 29 21:53:25 crc kubenswrapper[4741]: I0929 21:53:25.603135 4741 generic.go:334] "Generic (PLEG): container finished" podID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerID="de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12" exitCode=0 Sep 29 21:53:25 crc kubenswrapper[4741]: I0929 21:53:25.603173 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chpfj" event={"ID":"fdade596-ab6b-41ae-8c25-2f848ace7f87","Type":"ContainerDied","Data":"de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12"} Sep 29 21:53:25 crc kubenswrapper[4741]: I0929 21:53:25.603702 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chpfj" event={"ID":"fdade596-ab6b-41ae-8c25-2f848ace7f87","Type":"ContainerStarted","Data":"a96aa53245d0ed57eeb5916a4e3dac9a9eeff157e6c6adb11c1b55cc753c6bdb"} Sep 29 21:53:25 crc kubenswrapper[4741]: I0929 21:53:25.605689 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 21:53:25 crc kubenswrapper[4741]: I0929 21:53:25.675836 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Sep 29 21:53:25 crc kubenswrapper[4741]: I0929 21:53:25.676247 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="e16f6002-f781-4228-aaa2-458b21b72250" containerName="adoption" containerID="cri-o://092fffe8c0dcbe8b2043ef01bf0ffb3e1d9899414c05f285c2da79e8254f17d9" gracePeriod=30 Sep 29 21:53:27 crc kubenswrapper[4741]: I0929 21:53:27.086851 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:53:27 crc kubenswrapper[4741]: E0929 21:53:27.087342 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 21:53:27 crc kubenswrapper[4741]: I0929 21:53:27.622511 4741 generic.go:334] "Generic (PLEG): container finished" podID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerID="ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306" exitCode=0 Sep 29 21:53:27 crc kubenswrapper[4741]: I0929 21:53:27.622850 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chpfj" event={"ID":"fdade596-ab6b-41ae-8c25-2f848ace7f87","Type":"ContainerDied","Data":"ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306"} Sep 29 21:53:28 crc kubenswrapper[4741]: I0929 21:53:28.635477 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chpfj" event={"ID":"fdade596-ab6b-41ae-8c25-2f848ace7f87","Type":"ContainerStarted","Data":"f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa"} Sep 29 21:53:28 crc kubenswrapper[4741]: I0929 21:53:28.658459 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-chpfj" podStartSLOduration=2.155790692 podStartE2EDuration="4.658439015s" podCreationTimestamp="2025-09-29 21:53:24 +0000 UTC" firstStartedPulling="2025-09-29 21:53:25.605448439 +0000 UTC m=+9847.253237771" lastFinishedPulling="2025-09-29 21:53:28.108096762 +0000 UTC m=+9849.755886094" observedRunningTime="2025-09-29 21:53:28.653027385 +0000 UTC m=+9850.300816727" watchObservedRunningTime="2025-09-29 21:53:28.658439015 +0000 UTC m=+9850.306228347" Sep 29 21:53:34 crc kubenswrapper[4741]: I0929 21:53:34.591645 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:34 crc kubenswrapper[4741]: I0929 21:53:34.591962 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:34 crc kubenswrapper[4741]: I0929 21:53:34.639285 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:34 crc kubenswrapper[4741]: I0929 21:53:34.736792 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:34 crc kubenswrapper[4741]: I0929 21:53:34.879216 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-chpfj"] Sep 29 21:53:36 crc kubenswrapper[4741]: I0929 21:53:36.716706 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-chpfj" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="registry-server" containerID="cri-o://f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa" gracePeriod=2 Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.254933 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.397753 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpkzm\" (UniqueName: \"kubernetes.io/projected/fdade596-ab6b-41ae-8c25-2f848ace7f87-kube-api-access-lpkzm\") pod \"fdade596-ab6b-41ae-8c25-2f848ace7f87\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.397985 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-utilities\") pod \"fdade596-ab6b-41ae-8c25-2f848ace7f87\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.398148 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-catalog-content\") pod \"fdade596-ab6b-41ae-8c25-2f848ace7f87\" (UID: \"fdade596-ab6b-41ae-8c25-2f848ace7f87\") " Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.400127 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-utilities" (OuterVolumeSpecName: "utilities") pod "fdade596-ab6b-41ae-8c25-2f848ace7f87" (UID: "fdade596-ab6b-41ae-8c25-2f848ace7f87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.407550 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdade596-ab6b-41ae-8c25-2f848ace7f87-kube-api-access-lpkzm" (OuterVolumeSpecName: "kube-api-access-lpkzm") pod "fdade596-ab6b-41ae-8c25-2f848ace7f87" (UID: "fdade596-ab6b-41ae-8c25-2f848ace7f87"). InnerVolumeSpecName "kube-api-access-lpkzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.411724 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdade596-ab6b-41ae-8c25-2f848ace7f87" (UID: "fdade596-ab6b-41ae-8c25-2f848ace7f87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.499922 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpkzm\" (UniqueName: \"kubernetes.io/projected/fdade596-ab6b-41ae-8c25-2f848ace7f87-kube-api-access-lpkzm\") on node \"crc\" DevicePath \"\"" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.499955 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.499964 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdade596-ab6b-41ae-8c25-2f848ace7f87-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.728232 4741 generic.go:334] "Generic (PLEG): container finished" podID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerID="f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa" exitCode=0 Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.728283 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chpfj" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.728308 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chpfj" event={"ID":"fdade596-ab6b-41ae-8c25-2f848ace7f87","Type":"ContainerDied","Data":"f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa"} Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.728803 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chpfj" event={"ID":"fdade596-ab6b-41ae-8c25-2f848ace7f87","Type":"ContainerDied","Data":"a96aa53245d0ed57eeb5916a4e3dac9a9eeff157e6c6adb11c1b55cc753c6bdb"} Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.728830 4741 scope.go:117] "RemoveContainer" containerID="f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.748435 4741 scope.go:117] "RemoveContainer" containerID="ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.772614 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-chpfj"] Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.797942 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-chpfj"] Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.817784 4741 scope.go:117] "RemoveContainer" containerID="de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.866702 4741 scope.go:117] "RemoveContainer" containerID="f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa" Sep 29 21:53:37 crc kubenswrapper[4741]: E0929 21:53:37.867791 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa\": container with ID starting with f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa not found: ID does not exist" containerID="f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.867838 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa"} err="failed to get container status \"f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa\": rpc error: code = NotFound desc = could not find container \"f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa\": container with ID starting with f93db580ebcd940c53a088d513351e1e7c74e60e9f18828fd5cfeca53e8398aa not found: ID does not exist" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.867863 4741 scope.go:117] "RemoveContainer" containerID="ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306" Sep 29 21:53:37 crc kubenswrapper[4741]: E0929 21:53:37.868172 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306\": container with ID starting with ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306 not found: ID does not exist" containerID="ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.868216 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306"} err="failed to get container status \"ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306\": rpc error: code = NotFound desc = could not find container \"ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306\": container with ID starting with ddf5dd2f4c9c7f18e1821c9ea45c34b6227ec26df4d522413dda9a14e0d2d306 not found: ID does not exist" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.868243 4741 scope.go:117] "RemoveContainer" containerID="de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12" Sep 29 21:53:37 crc kubenswrapper[4741]: E0929 21:53:37.868582 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12\": container with ID starting with de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12 not found: ID does not exist" containerID="de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12" Sep 29 21:53:37 crc kubenswrapper[4741]: I0929 21:53:37.868612 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12"} err="failed to get container status \"de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12\": rpc error: code = NotFound desc = could not find container \"de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12\": container with ID starting with de3a4b10d5c46d1f92e19120592be0d78d08fef4c0d771133ad5a276e1136f12 not found: ID does not exist" Sep 29 21:53:39 crc kubenswrapper[4741]: I0929 21:53:39.092595 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:53:39 crc kubenswrapper[4741]: I0929 21:53:39.110152 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" path="/var/lib/kubelet/pods/fdade596-ab6b-41ae-8c25-2f848ace7f87/volumes" Sep 29 21:53:39 crc kubenswrapper[4741]: I0929 21:53:39.752639 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"fdc380dcd36ba60be6b9a434fe90ce73ff08cd0c48d301bfb2e40232d9aeccef"} Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.005606 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zv7jd"] Sep 29 21:53:55 crc kubenswrapper[4741]: E0929 21:53:55.006877 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="extract-utilities" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.006903 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="extract-utilities" Sep 29 21:53:55 crc kubenswrapper[4741]: E0929 21:53:55.006935 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="registry-server" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.006945 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="registry-server" Sep 29 21:53:55 crc kubenswrapper[4741]: E0929 21:53:55.006964 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="extract-content" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.006971 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="extract-content" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.007285 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdade596-ab6b-41ae-8c25-2f848ace7f87" containerName="registry-server" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.009265 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.017536 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zv7jd"] Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.168172 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfjm4\" (UniqueName: \"kubernetes.io/projected/a1448782-b74d-488c-82fd-7302b5483526-kube-api-access-vfjm4\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.168515 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-catalog-content\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.168708 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-utilities\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.271001 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-catalog-content\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.271103 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-utilities\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.271249 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfjm4\" (UniqueName: \"kubernetes.io/projected/a1448782-b74d-488c-82fd-7302b5483526-kube-api-access-vfjm4\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.271555 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-catalog-content\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.271911 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-utilities\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.294829 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfjm4\" (UniqueName: \"kubernetes.io/projected/a1448782-b74d-488c-82fd-7302b5483526-kube-api-access-vfjm4\") pod \"redhat-operators-zv7jd\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.341581 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.822486 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zv7jd"] Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.937431 4741 generic.go:334] "Generic (PLEG): container finished" podID="e16f6002-f781-4228-aaa2-458b21b72250" containerID="092fffe8c0dcbe8b2043ef01bf0ffb3e1d9899414c05f285c2da79e8254f17d9" exitCode=137 Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.937517 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"e16f6002-f781-4228-aaa2-458b21b72250","Type":"ContainerDied","Data":"092fffe8c0dcbe8b2043ef01bf0ffb3e1d9899414c05f285c2da79e8254f17d9"} Sep 29 21:53:55 crc kubenswrapper[4741]: I0929 21:53:55.944411 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv7jd" event={"ID":"a1448782-b74d-488c-82fd-7302b5483526","Type":"ContainerStarted","Data":"c806b7bced2cb17253214aff4cc5278bc511f32f19a423393581274d72770bb7"} Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.124259 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.294234 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\") pod \"e16f6002-f781-4228-aaa2-458b21b72250\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.294492 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnjbg\" (UniqueName: \"kubernetes.io/projected/e16f6002-f781-4228-aaa2-458b21b72250-kube-api-access-lnjbg\") pod \"e16f6002-f781-4228-aaa2-458b21b72250\" (UID: \"e16f6002-f781-4228-aaa2-458b21b72250\") " Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.303592 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e16f6002-f781-4228-aaa2-458b21b72250-kube-api-access-lnjbg" (OuterVolumeSpecName: "kube-api-access-lnjbg") pod "e16f6002-f781-4228-aaa2-458b21b72250" (UID: "e16f6002-f781-4228-aaa2-458b21b72250"). InnerVolumeSpecName "kube-api-access-lnjbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.317208 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8" (OuterVolumeSpecName: "mariadb-data") pod "e16f6002-f781-4228-aaa2-458b21b72250" (UID: "e16f6002-f781-4228-aaa2-458b21b72250"). InnerVolumeSpecName "pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.397545 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnjbg\" (UniqueName: \"kubernetes.io/projected/e16f6002-f781-4228-aaa2-458b21b72250-kube-api-access-lnjbg\") on node \"crc\" DevicePath \"\"" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.397628 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\") on node \"crc\" " Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.424939 4741 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.425344 4741 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8") on node "crc" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.499105 4741 reconciler_common.go:293] "Volume detached for volume \"pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c2be7de-7119-400a-8c62-ac4eb86b01b8\") on node \"crc\" DevicePath \"\"" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.957078 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"e16f6002-f781-4228-aaa2-458b21b72250","Type":"ContainerDied","Data":"af8d87447c1239e3803c6476090dabc12c4ab76b97a01288dafe6840926dffcf"} Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.957117 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.957141 4741 scope.go:117] "RemoveContainer" containerID="092fffe8c0dcbe8b2043ef01bf0ffb3e1d9899414c05f285c2da79e8254f17d9" Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.959852 4741 generic.go:334] "Generic (PLEG): container finished" podID="a1448782-b74d-488c-82fd-7302b5483526" containerID="1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3" exitCode=0 Sep 29 21:53:56 crc kubenswrapper[4741]: I0929 21:53:56.959892 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv7jd" event={"ID":"a1448782-b74d-488c-82fd-7302b5483526","Type":"ContainerDied","Data":"1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3"} Sep 29 21:53:57 crc kubenswrapper[4741]: I0929 21:53:57.008474 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Sep 29 21:53:57 crc kubenswrapper[4741]: I0929 21:53:57.017222 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Sep 29 21:53:57 crc kubenswrapper[4741]: I0929 21:53:57.097523 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e16f6002-f781-4228-aaa2-458b21b72250" path="/var/lib/kubelet/pods/e16f6002-f781-4228-aaa2-458b21b72250/volumes" Sep 29 21:53:57 crc kubenswrapper[4741]: I0929 21:53:57.714562 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Sep 29 21:53:57 crc kubenswrapper[4741]: I0929 21:53:57.714850 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="39cc47ab-05f6-4582-aabc-71217551efc7" containerName="adoption" containerID="cri-o://c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4" gracePeriod=30 Sep 29 21:53:58 crc kubenswrapper[4741]: I0929 21:53:58.984233 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv7jd" event={"ID":"a1448782-b74d-488c-82fd-7302b5483526","Type":"ContainerStarted","Data":"58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29"} Sep 29 21:53:59 crc kubenswrapper[4741]: I0929 21:53:59.997220 4741 generic.go:334] "Generic (PLEG): container finished" podID="a1448782-b74d-488c-82fd-7302b5483526" containerID="58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29" exitCode=0 Sep 29 21:53:59 crc kubenswrapper[4741]: I0929 21:53:59.997323 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv7jd" event={"ID":"a1448782-b74d-488c-82fd-7302b5483526","Type":"ContainerDied","Data":"58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29"} Sep 29 21:54:02 crc kubenswrapper[4741]: I0929 21:54:02.020175 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv7jd" event={"ID":"a1448782-b74d-488c-82fd-7302b5483526","Type":"ContainerStarted","Data":"7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539"} Sep 29 21:54:02 crc kubenswrapper[4741]: I0929 21:54:02.048917 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zv7jd" podStartSLOduration=4.527515541 podStartE2EDuration="8.048896716s" podCreationTimestamp="2025-09-29 21:53:54 +0000 UTC" firstStartedPulling="2025-09-29 21:53:56.961849982 +0000 UTC m=+9878.609639314" lastFinishedPulling="2025-09-29 21:54:00.483231157 +0000 UTC m=+9882.131020489" observedRunningTime="2025-09-29 21:54:02.040383008 +0000 UTC m=+9883.688172340" watchObservedRunningTime="2025-09-29 21:54:02.048896716 +0000 UTC m=+9883.696686048" Sep 29 21:54:05 crc kubenswrapper[4741]: I0929 21:54:05.342544 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:54:05 crc kubenswrapper[4741]: I0929 21:54:05.343182 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:54:05 crc kubenswrapper[4741]: I0929 21:54:05.392233 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:54:06 crc kubenswrapper[4741]: I0929 21:54:06.098553 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:54:06 crc kubenswrapper[4741]: I0929 21:54:06.140822 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zv7jd"] Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.074168 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zv7jd" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="registry-server" containerID="cri-o://7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539" gracePeriod=2 Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.622612 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.787283 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-catalog-content\") pod \"a1448782-b74d-488c-82fd-7302b5483526\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.787496 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfjm4\" (UniqueName: \"kubernetes.io/projected/a1448782-b74d-488c-82fd-7302b5483526-kube-api-access-vfjm4\") pod \"a1448782-b74d-488c-82fd-7302b5483526\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.787568 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-utilities\") pod \"a1448782-b74d-488c-82fd-7302b5483526\" (UID: \"a1448782-b74d-488c-82fd-7302b5483526\") " Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.795135 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1448782-b74d-488c-82fd-7302b5483526-kube-api-access-vfjm4" (OuterVolumeSpecName: "kube-api-access-vfjm4") pod "a1448782-b74d-488c-82fd-7302b5483526" (UID: "a1448782-b74d-488c-82fd-7302b5483526"). InnerVolumeSpecName "kube-api-access-vfjm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.797672 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-utilities" (OuterVolumeSpecName: "utilities") pod "a1448782-b74d-488c-82fd-7302b5483526" (UID: "a1448782-b74d-488c-82fd-7302b5483526"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.875700 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1448782-b74d-488c-82fd-7302b5483526" (UID: "a1448782-b74d-488c-82fd-7302b5483526"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.890084 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.890122 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfjm4\" (UniqueName: \"kubernetes.io/projected/a1448782-b74d-488c-82fd-7302b5483526-kube-api-access-vfjm4\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:08 crc kubenswrapper[4741]: I0929 21:54:08.890134 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1448782-b74d-488c-82fd-7302b5483526-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.085499 4741 generic.go:334] "Generic (PLEG): container finished" podID="a1448782-b74d-488c-82fd-7302b5483526" containerID="7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539" exitCode=0 Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.091821 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv7jd" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.101370 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv7jd" event={"ID":"a1448782-b74d-488c-82fd-7302b5483526","Type":"ContainerDied","Data":"7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539"} Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.101436 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv7jd" event={"ID":"a1448782-b74d-488c-82fd-7302b5483526","Type":"ContainerDied","Data":"c806b7bced2cb17253214aff4cc5278bc511f32f19a423393581274d72770bb7"} Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.101456 4741 scope.go:117] "RemoveContainer" containerID="7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.128609 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zv7jd"] Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.133635 4741 scope.go:117] "RemoveContainer" containerID="58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.140570 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zv7jd"] Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.608354 4741 scope.go:117] "RemoveContainer" containerID="1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.658307 4741 scope.go:117] "RemoveContainer" containerID="7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539" Sep 29 21:54:09 crc kubenswrapper[4741]: E0929 21:54:09.658837 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539\": container with ID starting with 7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539 not found: ID does not exist" containerID="7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.658880 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539"} err="failed to get container status \"7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539\": rpc error: code = NotFound desc = could not find container \"7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539\": container with ID starting with 7702130ced70335e20fb7b835feae376040e5baf675a3d0d595b54bfcc65d539 not found: ID does not exist" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.658916 4741 scope.go:117] "RemoveContainer" containerID="58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29" Sep 29 21:54:09 crc kubenswrapper[4741]: E0929 21:54:09.659274 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29\": container with ID starting with 58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29 not found: ID does not exist" containerID="58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.659323 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29"} err="failed to get container status \"58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29\": rpc error: code = NotFound desc = could not find container \"58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29\": container with ID starting with 58fec43a6159bc10d81a6a38d0848b9adfe8dece58d2eb344def62cb48f8cd29 not found: ID does not exist" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.659359 4741 scope.go:117] "RemoveContainer" containerID="1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3" Sep 29 21:54:09 crc kubenswrapper[4741]: E0929 21:54:09.659719 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3\": container with ID starting with 1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3 not found: ID does not exist" containerID="1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3" Sep 29 21:54:09 crc kubenswrapper[4741]: I0929 21:54:09.659747 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3"} err="failed to get container status \"1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3\": rpc error: code = NotFound desc = could not find container \"1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3\": container with ID starting with 1542831209f2a3111ac0b55f49b6fd3ad6ba995e206c35d3d642218ac84b67d3 not found: ID does not exist" Sep 29 21:54:11 crc kubenswrapper[4741]: I0929 21:54:11.096881 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1448782-b74d-488c-82fd-7302b5483526" path="/var/lib/kubelet/pods/a1448782-b74d-488c-82fd-7302b5483526/volumes" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.736101 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j2nm5"] Sep 29 21:54:18 crc kubenswrapper[4741]: E0929 21:54:18.737218 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="registry-server" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.737233 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="registry-server" Sep 29 21:54:18 crc kubenswrapper[4741]: E0929 21:54:18.737255 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e16f6002-f781-4228-aaa2-458b21b72250" containerName="adoption" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.737263 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="e16f6002-f781-4228-aaa2-458b21b72250" containerName="adoption" Sep 29 21:54:18 crc kubenswrapper[4741]: E0929 21:54:18.737293 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="extract-utilities" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.737299 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="extract-utilities" Sep 29 21:54:18 crc kubenswrapper[4741]: E0929 21:54:18.737317 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="extract-content" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.737323 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="extract-content" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.737808 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1448782-b74d-488c-82fd-7302b5483526" containerName="registry-server" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.737834 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="e16f6002-f781-4228-aaa2-458b21b72250" containerName="adoption" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.741031 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.762664 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j2nm5"] Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.888898 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ldxh\" (UniqueName: \"kubernetes.io/projected/c3064e23-02d1-43f7-bd27-2664c410079c-kube-api-access-2ldxh\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.888947 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-catalog-content\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.889003 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-utilities\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.991359 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ldxh\" (UniqueName: \"kubernetes.io/projected/c3064e23-02d1-43f7-bd27-2664c410079c-kube-api-access-2ldxh\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.991879 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-catalog-content\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.992021 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-utilities\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.992420 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-catalog-content\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:18 crc kubenswrapper[4741]: I0929 21:54:18.992450 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-utilities\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:19 crc kubenswrapper[4741]: I0929 21:54:19.503191 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ldxh\" (UniqueName: \"kubernetes.io/projected/c3064e23-02d1-43f7-bd27-2664c410079c-kube-api-access-2ldxh\") pod \"certified-operators-j2nm5\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:19 crc kubenswrapper[4741]: I0929 21:54:19.680060 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:20 crc kubenswrapper[4741]: I0929 21:54:20.196374 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j2nm5"] Sep 29 21:54:20 crc kubenswrapper[4741]: W0929 21:54:20.205666 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3064e23_02d1_43f7_bd27_2664c410079c.slice/crio-d64305cd9c1c18e4345375d63235d70cf03ee5e69e23f4b5cdf70e42b4e75027 WatchSource:0}: Error finding container d64305cd9c1c18e4345375d63235d70cf03ee5e69e23f4b5cdf70e42b4e75027: Status 404 returned error can't find the container with id d64305cd9c1c18e4345375d63235d70cf03ee5e69e23f4b5cdf70e42b4e75027 Sep 29 21:54:21 crc kubenswrapper[4741]: I0929 21:54:21.229606 4741 generic.go:334] "Generic (PLEG): container finished" podID="c3064e23-02d1-43f7-bd27-2664c410079c" containerID="a98b3da73304b6a34b133c3c129b8da04478e2a94311d056c796256561d93fd1" exitCode=0 Sep 29 21:54:21 crc kubenswrapper[4741]: I0929 21:54:21.229723 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nm5" event={"ID":"c3064e23-02d1-43f7-bd27-2664c410079c","Type":"ContainerDied","Data":"a98b3da73304b6a34b133c3c129b8da04478e2a94311d056c796256561d93fd1"} Sep 29 21:54:21 crc kubenswrapper[4741]: I0929 21:54:21.230043 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nm5" event={"ID":"c3064e23-02d1-43f7-bd27-2664c410079c","Type":"ContainerStarted","Data":"d64305cd9c1c18e4345375d63235d70cf03ee5e69e23f4b5cdf70e42b4e75027"} Sep 29 21:54:23 crc kubenswrapper[4741]: I0929 21:54:23.252050 4741 generic.go:334] "Generic (PLEG): container finished" podID="c3064e23-02d1-43f7-bd27-2664c410079c" containerID="41e282d53c8e309d8403b3ab211d95d7865e7bc4d27ed2056f3aad23b7f96dc7" exitCode=0 Sep 29 21:54:23 crc kubenswrapper[4741]: I0929 21:54:23.252145 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nm5" event={"ID":"c3064e23-02d1-43f7-bd27-2664c410079c","Type":"ContainerDied","Data":"41e282d53c8e309d8403b3ab211d95d7865e7bc4d27ed2056f3aad23b7f96dc7"} Sep 29 21:54:24 crc kubenswrapper[4741]: I0929 21:54:24.263639 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nm5" event={"ID":"c3064e23-02d1-43f7-bd27-2664c410079c","Type":"ContainerStarted","Data":"d4ab41e65c9d70f58cbcf29e1ce7b89841f760453ef914414d36bd48e25d863c"} Sep 29 21:54:24 crc kubenswrapper[4741]: I0929 21:54:24.283667 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j2nm5" podStartSLOduration=3.84177249 podStartE2EDuration="6.283648384s" podCreationTimestamp="2025-09-29 21:54:18 +0000 UTC" firstStartedPulling="2025-09-29 21:54:21.23229368 +0000 UTC m=+9902.880083012" lastFinishedPulling="2025-09-29 21:54:23.674169574 +0000 UTC m=+9905.321958906" observedRunningTime="2025-09-29 21:54:24.278961827 +0000 UTC m=+9905.926751159" watchObservedRunningTime="2025-09-29 21:54:24.283648384 +0000 UTC m=+9905.931437716" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.282875 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.316206 4741 generic.go:334] "Generic (PLEG): container finished" podID="39cc47ab-05f6-4582-aabc-71217551efc7" containerID="c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4" exitCode=137 Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.316255 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"39cc47ab-05f6-4582-aabc-71217551efc7","Type":"ContainerDied","Data":"c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4"} Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.316286 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"39cc47ab-05f6-4582-aabc-71217551efc7","Type":"ContainerDied","Data":"8e9ba2fd429bfabdfb6e3188196d71ee543cc555aa407eb54a29ca91bd57df31"} Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.316303 4741 scope.go:117] "RemoveContainer" containerID="c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.316469 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.343167 4741 scope.go:117] "RemoveContainer" containerID="c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4" Sep 29 21:54:28 crc kubenswrapper[4741]: E0929 21:54:28.343540 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4\": container with ID starting with c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4 not found: ID does not exist" containerID="c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.343585 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4"} err="failed to get container status \"c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4\": rpc error: code = NotFound desc = could not find container \"c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4\": container with ID starting with c1574225577bb4570a35049923de8f2fad00d331515fe30d4b79d8b2d2724fd4 not found: ID does not exist" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.457778 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286\") pod \"39cc47ab-05f6-4582-aabc-71217551efc7\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.458344 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/39cc47ab-05f6-4582-aabc-71217551efc7-ovn-data-cert\") pod \"39cc47ab-05f6-4582-aabc-71217551efc7\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.458441 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n2bg\" (UniqueName: \"kubernetes.io/projected/39cc47ab-05f6-4582-aabc-71217551efc7-kube-api-access-2n2bg\") pod \"39cc47ab-05f6-4582-aabc-71217551efc7\" (UID: \"39cc47ab-05f6-4582-aabc-71217551efc7\") " Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.465080 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cc47ab-05f6-4582-aabc-71217551efc7-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "39cc47ab-05f6-4582-aabc-71217551efc7" (UID: "39cc47ab-05f6-4582-aabc-71217551efc7"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.465651 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39cc47ab-05f6-4582-aabc-71217551efc7-kube-api-access-2n2bg" (OuterVolumeSpecName: "kube-api-access-2n2bg") pod "39cc47ab-05f6-4582-aabc-71217551efc7" (UID: "39cc47ab-05f6-4582-aabc-71217551efc7"). InnerVolumeSpecName "kube-api-access-2n2bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.479843 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286" (OuterVolumeSpecName: "ovn-data") pod "39cc47ab-05f6-4582-aabc-71217551efc7" (UID: "39cc47ab-05f6-4582-aabc-71217551efc7"). InnerVolumeSpecName "pvc-41089285-8dcf-44b4-8a50-2286ad169286". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.561844 4741 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/39cc47ab-05f6-4582-aabc-71217551efc7-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.561885 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n2bg\" (UniqueName: \"kubernetes.io/projected/39cc47ab-05f6-4582-aabc-71217551efc7-kube-api-access-2n2bg\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.561931 4741 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-41089285-8dcf-44b4-8a50-2286ad169286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286\") on node \"crc\" " Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.589750 4741 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.589919 4741 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-41089285-8dcf-44b4-8a50-2286ad169286" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286") on node "crc" Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.652652 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.662343 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Sep 29 21:54:28 crc kubenswrapper[4741]: I0929 21:54:28.663582 4741 reconciler_common.go:293] "Volume detached for volume \"pvc-41089285-8dcf-44b4-8a50-2286ad169286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-41089285-8dcf-44b4-8a50-2286ad169286\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:29 crc kubenswrapper[4741]: I0929 21:54:29.099693 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39cc47ab-05f6-4582-aabc-71217551efc7" path="/var/lib/kubelet/pods/39cc47ab-05f6-4582-aabc-71217551efc7/volumes" Sep 29 21:54:29 crc kubenswrapper[4741]: I0929 21:54:29.680575 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:29 crc kubenswrapper[4741]: I0929 21:54:29.681071 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:29 crc kubenswrapper[4741]: I0929 21:54:29.753536 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:30 crc kubenswrapper[4741]: I0929 21:54:30.750987 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:30 crc kubenswrapper[4741]: I0929 21:54:30.805271 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j2nm5"] Sep 29 21:54:32 crc kubenswrapper[4741]: I0929 21:54:32.365663 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j2nm5" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="registry-server" containerID="cri-o://d4ab41e65c9d70f58cbcf29e1ce7b89841f760453ef914414d36bd48e25d863c" gracePeriod=2 Sep 29 21:54:33 crc kubenswrapper[4741]: I0929 21:54:33.377703 4741 generic.go:334] "Generic (PLEG): container finished" podID="c3064e23-02d1-43f7-bd27-2664c410079c" containerID="d4ab41e65c9d70f58cbcf29e1ce7b89841f760453ef914414d36bd48e25d863c" exitCode=0 Sep 29 21:54:33 crc kubenswrapper[4741]: I0929 21:54:33.377782 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nm5" event={"ID":"c3064e23-02d1-43f7-bd27-2664c410079c","Type":"ContainerDied","Data":"d4ab41e65c9d70f58cbcf29e1ce7b89841f760453ef914414d36bd48e25d863c"} Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.007166 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.101027 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-catalog-content\") pod \"c3064e23-02d1-43f7-bd27-2664c410079c\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.101099 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-utilities\") pod \"c3064e23-02d1-43f7-bd27-2664c410079c\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.101131 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ldxh\" (UniqueName: \"kubernetes.io/projected/c3064e23-02d1-43f7-bd27-2664c410079c-kube-api-access-2ldxh\") pod \"c3064e23-02d1-43f7-bd27-2664c410079c\" (UID: \"c3064e23-02d1-43f7-bd27-2664c410079c\") " Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.103454 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-utilities" (OuterVolumeSpecName: "utilities") pod "c3064e23-02d1-43f7-bd27-2664c410079c" (UID: "c3064e23-02d1-43f7-bd27-2664c410079c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.130795 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3064e23-02d1-43f7-bd27-2664c410079c-kube-api-access-2ldxh" (OuterVolumeSpecName: "kube-api-access-2ldxh") pod "c3064e23-02d1-43f7-bd27-2664c410079c" (UID: "c3064e23-02d1-43f7-bd27-2664c410079c"). InnerVolumeSpecName "kube-api-access-2ldxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.221454 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.221497 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ldxh\" (UniqueName: \"kubernetes.io/projected/c3064e23-02d1-43f7-bd27-2664c410079c-kube-api-access-2ldxh\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.243944 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3064e23-02d1-43f7-bd27-2664c410079c" (UID: "c3064e23-02d1-43f7-bd27-2664c410079c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.324995 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3064e23-02d1-43f7-bd27-2664c410079c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.391051 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nm5" event={"ID":"c3064e23-02d1-43f7-bd27-2664c410079c","Type":"ContainerDied","Data":"d64305cd9c1c18e4345375d63235d70cf03ee5e69e23f4b5cdf70e42b4e75027"} Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.391122 4741 scope.go:117] "RemoveContainer" containerID="d4ab41e65c9d70f58cbcf29e1ce7b89841f760453ef914414d36bd48e25d863c" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.391128 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nm5" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.420897 4741 scope.go:117] "RemoveContainer" containerID="41e282d53c8e309d8403b3ab211d95d7865e7bc4d27ed2056f3aad23b7f96dc7" Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.437177 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j2nm5"] Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.448594 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j2nm5"] Sep 29 21:54:34 crc kubenswrapper[4741]: I0929 21:54:34.453923 4741 scope.go:117] "RemoveContainer" containerID="a98b3da73304b6a34b133c3c129b8da04478e2a94311d056c796256561d93fd1" Sep 29 21:54:35 crc kubenswrapper[4741]: I0929 21:54:35.116621 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" path="/var/lib/kubelet/pods/c3064e23-02d1-43f7-bd27-2664c410079c/volumes" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.823280 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4m44r/must-gather-dqv2x"] Sep 29 21:55:32 crc kubenswrapper[4741]: E0929 21:55:32.824182 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="extract-content" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.824197 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="extract-content" Sep 29 21:55:32 crc kubenswrapper[4741]: E0929 21:55:32.824222 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="registry-server" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.824228 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="registry-server" Sep 29 21:55:32 crc kubenswrapper[4741]: E0929 21:55:32.824240 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="extract-utilities" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.824247 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="extract-utilities" Sep 29 21:55:32 crc kubenswrapper[4741]: E0929 21:55:32.824269 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cc47ab-05f6-4582-aabc-71217551efc7" containerName="adoption" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.824274 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cc47ab-05f6-4582-aabc-71217551efc7" containerName="adoption" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.824489 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cc47ab-05f6-4582-aabc-71217551efc7" containerName="adoption" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.824511 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3064e23-02d1-43f7-bd27-2664c410079c" containerName="registry-server" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.825655 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.858424 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4m44r"/"openshift-service-ca.crt" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.858712 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4m44r"/"kube-root-ca.crt" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.858903 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4m44r"/"default-dockercfg-nkcbq" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.868963 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4m44r/must-gather-dqv2x"] Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.915501 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xk9c\" (UniqueName: \"kubernetes.io/projected/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-kube-api-access-4xk9c\") pod \"must-gather-dqv2x\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:32 crc kubenswrapper[4741]: I0929 21:55:32.915566 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-must-gather-output\") pod \"must-gather-dqv2x\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:33 crc kubenswrapper[4741]: I0929 21:55:33.017218 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xk9c\" (UniqueName: \"kubernetes.io/projected/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-kube-api-access-4xk9c\") pod \"must-gather-dqv2x\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:33 crc kubenswrapper[4741]: I0929 21:55:33.017288 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-must-gather-output\") pod \"must-gather-dqv2x\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:33 crc kubenswrapper[4741]: I0929 21:55:33.017886 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-must-gather-output\") pod \"must-gather-dqv2x\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:33 crc kubenswrapper[4741]: I0929 21:55:33.036076 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xk9c\" (UniqueName: \"kubernetes.io/projected/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-kube-api-access-4xk9c\") pod \"must-gather-dqv2x\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:33 crc kubenswrapper[4741]: I0929 21:55:33.177894 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 21:55:33 crc kubenswrapper[4741]: I0929 21:55:33.687045 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4m44r/must-gather-dqv2x"] Sep 29 21:55:33 crc kubenswrapper[4741]: I0929 21:55:33.994304 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/must-gather-dqv2x" event={"ID":"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7","Type":"ContainerStarted","Data":"3cf8981a349fdda3c3ebe9d71249d5bb7b03be2a2836cff384cce9d5a72fc037"} Sep 29 21:55:38 crc kubenswrapper[4741]: I0929 21:55:38.039166 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/must-gather-dqv2x" event={"ID":"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7","Type":"ContainerStarted","Data":"7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260"} Sep 29 21:55:39 crc kubenswrapper[4741]: I0929 21:55:39.052903 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/must-gather-dqv2x" event={"ID":"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7","Type":"ContainerStarted","Data":"116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32"} Sep 29 21:55:39 crc kubenswrapper[4741]: I0929 21:55:39.074691 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4m44r/must-gather-dqv2x" podStartSLOduration=2.996775169 podStartE2EDuration="7.074672491s" podCreationTimestamp="2025-09-29 21:55:32 +0000 UTC" firstStartedPulling="2025-09-29 21:55:33.681959799 +0000 UTC m=+9975.329749131" lastFinishedPulling="2025-09-29 21:55:37.759857121 +0000 UTC m=+9979.407646453" observedRunningTime="2025-09-29 21:55:39.071869283 +0000 UTC m=+9980.719658615" watchObservedRunningTime="2025-09-29 21:55:39.074672491 +0000 UTC m=+9980.722461823" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.187927 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4m44r/crc-debug-pqhls"] Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.190357 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.217800 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cd3776d-f1db-4f09-a62e-3ffc7a681187-host\") pod \"crc-debug-pqhls\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.217979 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpg2x\" (UniqueName: \"kubernetes.io/projected/1cd3776d-f1db-4f09-a62e-3ffc7a681187-kube-api-access-jpg2x\") pod \"crc-debug-pqhls\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.319736 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cd3776d-f1db-4f09-a62e-3ffc7a681187-host\") pod \"crc-debug-pqhls\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.319903 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cd3776d-f1db-4f09-a62e-3ffc7a681187-host\") pod \"crc-debug-pqhls\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.319992 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpg2x\" (UniqueName: \"kubernetes.io/projected/1cd3776d-f1db-4f09-a62e-3ffc7a681187-kube-api-access-jpg2x\") pod \"crc-debug-pqhls\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.338052 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpg2x\" (UniqueName: \"kubernetes.io/projected/1cd3776d-f1db-4f09-a62e-3ffc7a681187-kube-api-access-jpg2x\") pod \"crc-debug-pqhls\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:42 crc kubenswrapper[4741]: I0929 21:55:42.510604 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:55:43 crc kubenswrapper[4741]: I0929 21:55:43.096078 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-pqhls" event={"ID":"1cd3776d-f1db-4f09-a62e-3ffc7a681187","Type":"ContainerStarted","Data":"696cf9dbc44e1097d3c0ad5c48f1ef738139e79471b78962d0e1f52ef13ad6ac"} Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.392679 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5vljw"] Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.395878 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.410098 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5vljw"] Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.547912 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzsn9\" (UniqueName: \"kubernetes.io/projected/c37c96d5-cca1-4f84-a834-55944b188eda-kube-api-access-fzsn9\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.547990 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-utilities\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.548083 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-catalog-content\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.650018 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzsn9\" (UniqueName: \"kubernetes.io/projected/c37c96d5-cca1-4f84-a834-55944b188eda-kube-api-access-fzsn9\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.650390 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-utilities\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.650454 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-catalog-content\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.650913 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-utilities\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:48 crc kubenswrapper[4741]: I0929 21:55:48.651059 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-catalog-content\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:49 crc kubenswrapper[4741]: I0929 21:55:49.095084 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzsn9\" (UniqueName: \"kubernetes.io/projected/c37c96d5-cca1-4f84-a834-55944b188eda-kube-api-access-fzsn9\") pod \"community-operators-5vljw\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:49 crc kubenswrapper[4741]: I0929 21:55:49.342443 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:55 crc kubenswrapper[4741]: I0929 21:55:55.266551 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-pqhls" event={"ID":"1cd3776d-f1db-4f09-a62e-3ffc7a681187","Type":"ContainerStarted","Data":"d3e60fb8e9ffa9895a8f9f9932b18b3fd6ad6c0b655dcb0ddcac2c63545f8b8d"} Sep 29 21:55:55 crc kubenswrapper[4741]: I0929 21:55:55.289871 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4m44r/crc-debug-pqhls" podStartSLOduration=1.012713612 podStartE2EDuration="13.289851521s" podCreationTimestamp="2025-09-29 21:55:42 +0000 UTC" firstStartedPulling="2025-09-29 21:55:42.549891677 +0000 UTC m=+9984.197681009" lastFinishedPulling="2025-09-29 21:55:54.827029586 +0000 UTC m=+9996.474818918" observedRunningTime="2025-09-29 21:55:55.288194959 +0000 UTC m=+9996.935984291" watchObservedRunningTime="2025-09-29 21:55:55.289851521 +0000 UTC m=+9996.937640853" Sep 29 21:55:55 crc kubenswrapper[4741]: I0929 21:55:55.341481 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5vljw"] Sep 29 21:55:55 crc kubenswrapper[4741]: W0929 21:55:55.342241 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc37c96d5_cca1_4f84_a834_55944b188eda.slice/crio-a90431e74f1b621b0ab5a53431e29bc709708210e85fd60e42da757a1b499953 WatchSource:0}: Error finding container a90431e74f1b621b0ab5a53431e29bc709708210e85fd60e42da757a1b499953: Status 404 returned error can't find the container with id a90431e74f1b621b0ab5a53431e29bc709708210e85fd60e42da757a1b499953 Sep 29 21:55:56 crc kubenswrapper[4741]: I0929 21:55:56.278631 4741 generic.go:334] "Generic (PLEG): container finished" podID="c37c96d5-cca1-4f84-a834-55944b188eda" containerID="200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7" exitCode=0 Sep 29 21:55:56 crc kubenswrapper[4741]: I0929 21:55:56.278719 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vljw" event={"ID":"c37c96d5-cca1-4f84-a834-55944b188eda","Type":"ContainerDied","Data":"200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7"} Sep 29 21:55:56 crc kubenswrapper[4741]: I0929 21:55:56.280453 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vljw" event={"ID":"c37c96d5-cca1-4f84-a834-55944b188eda","Type":"ContainerStarted","Data":"a90431e74f1b621b0ab5a53431e29bc709708210e85fd60e42da757a1b499953"} Sep 29 21:55:57 crc kubenswrapper[4741]: I0929 21:55:57.291296 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vljw" event={"ID":"c37c96d5-cca1-4f84-a834-55944b188eda","Type":"ContainerStarted","Data":"47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd"} Sep 29 21:55:58 crc kubenswrapper[4741]: I0929 21:55:58.306105 4741 generic.go:334] "Generic (PLEG): container finished" podID="c37c96d5-cca1-4f84-a834-55944b188eda" containerID="47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd" exitCode=0 Sep 29 21:55:58 crc kubenswrapper[4741]: I0929 21:55:58.306190 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vljw" event={"ID":"c37c96d5-cca1-4f84-a834-55944b188eda","Type":"ContainerDied","Data":"47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd"} Sep 29 21:55:59 crc kubenswrapper[4741]: I0929 21:55:59.325330 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vljw" event={"ID":"c37c96d5-cca1-4f84-a834-55944b188eda","Type":"ContainerStarted","Data":"736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218"} Sep 29 21:55:59 crc kubenswrapper[4741]: I0929 21:55:59.344953 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:55:59 crc kubenswrapper[4741]: I0929 21:55:59.348437 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:56:00 crc kubenswrapper[4741]: I0929 21:56:00.402522 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-5vljw" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="registry-server" probeResult="failure" output=< Sep 29 21:56:00 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 21:56:00 crc kubenswrapper[4741]: > Sep 29 21:56:01 crc kubenswrapper[4741]: I0929 21:56:01.739319 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:56:01 crc kubenswrapper[4741]: I0929 21:56:01.739682 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:56:09 crc kubenswrapper[4741]: I0929 21:56:09.394494 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:56:09 crc kubenswrapper[4741]: I0929 21:56:09.423223 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5vljw" podStartSLOduration=18.874689541 podStartE2EDuration="21.423201104s" podCreationTimestamp="2025-09-29 21:55:48 +0000 UTC" firstStartedPulling="2025-09-29 21:55:56.281657498 +0000 UTC m=+9997.929446830" lastFinishedPulling="2025-09-29 21:55:58.830169061 +0000 UTC m=+10000.477958393" observedRunningTime="2025-09-29 21:55:59.344863944 +0000 UTC m=+10000.992653276" watchObservedRunningTime="2025-09-29 21:56:09.423201104 +0000 UTC m=+10011.070990436" Sep 29 21:56:09 crc kubenswrapper[4741]: I0929 21:56:09.452532 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:56:09 crc kubenswrapper[4741]: I0929 21:56:09.636720 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5vljw"] Sep 29 21:56:10 crc kubenswrapper[4741]: I0929 21:56:10.434021 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5vljw" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="registry-server" containerID="cri-o://736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218" gracePeriod=2 Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.031122 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.127339 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-catalog-content\") pod \"c37c96d5-cca1-4f84-a834-55944b188eda\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.127545 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-utilities\") pod \"c37c96d5-cca1-4f84-a834-55944b188eda\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.127613 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzsn9\" (UniqueName: \"kubernetes.io/projected/c37c96d5-cca1-4f84-a834-55944b188eda-kube-api-access-fzsn9\") pod \"c37c96d5-cca1-4f84-a834-55944b188eda\" (UID: \"c37c96d5-cca1-4f84-a834-55944b188eda\") " Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.129178 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-utilities" (OuterVolumeSpecName: "utilities") pod "c37c96d5-cca1-4f84-a834-55944b188eda" (UID: "c37c96d5-cca1-4f84-a834-55944b188eda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.140891 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c37c96d5-cca1-4f84-a834-55944b188eda-kube-api-access-fzsn9" (OuterVolumeSpecName: "kube-api-access-fzsn9") pod "c37c96d5-cca1-4f84-a834-55944b188eda" (UID: "c37c96d5-cca1-4f84-a834-55944b188eda"). InnerVolumeSpecName "kube-api-access-fzsn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.185881 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c37c96d5-cca1-4f84-a834-55944b188eda" (UID: "c37c96d5-cca1-4f84-a834-55944b188eda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.230237 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.230277 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37c96d5-cca1-4f84-a834-55944b188eda-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.230290 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzsn9\" (UniqueName: \"kubernetes.io/projected/c37c96d5-cca1-4f84-a834-55944b188eda-kube-api-access-fzsn9\") on node \"crc\" DevicePath \"\"" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.457560 4741 generic.go:334] "Generic (PLEG): container finished" podID="c37c96d5-cca1-4f84-a834-55944b188eda" containerID="736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218" exitCode=0 Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.457721 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vljw" event={"ID":"c37c96d5-cca1-4f84-a834-55944b188eda","Type":"ContainerDied","Data":"736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218"} Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.457920 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vljw" event={"ID":"c37c96d5-cca1-4f84-a834-55944b188eda","Type":"ContainerDied","Data":"a90431e74f1b621b0ab5a53431e29bc709708210e85fd60e42da757a1b499953"} Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.457951 4741 scope.go:117] "RemoveContainer" containerID="736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.457875 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vljw" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.502599 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5vljw"] Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.506778 4741 scope.go:117] "RemoveContainer" containerID="47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.512178 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5vljw"] Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.549656 4741 scope.go:117] "RemoveContainer" containerID="200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.594820 4741 scope.go:117] "RemoveContainer" containerID="736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218" Sep 29 21:56:11 crc kubenswrapper[4741]: E0929 21:56:11.595311 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218\": container with ID starting with 736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218 not found: ID does not exist" containerID="736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.595378 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218"} err="failed to get container status \"736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218\": rpc error: code = NotFound desc = could not find container \"736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218\": container with ID starting with 736cf0ad9bcace61dd93f5fe425d340459d2a046d5913475e1c36383d03bd218 not found: ID does not exist" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.595433 4741 scope.go:117] "RemoveContainer" containerID="47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd" Sep 29 21:56:11 crc kubenswrapper[4741]: E0929 21:56:11.595982 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd\": container with ID starting with 47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd not found: ID does not exist" containerID="47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.596038 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd"} err="failed to get container status \"47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd\": rpc error: code = NotFound desc = could not find container \"47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd\": container with ID starting with 47265eb0158b0e86b39529d0fa3ec0af31a4399feff7db37df92e591bc3947dd not found: ID does not exist" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.596065 4741 scope.go:117] "RemoveContainer" containerID="200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7" Sep 29 21:56:11 crc kubenswrapper[4741]: E0929 21:56:11.596476 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7\": container with ID starting with 200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7 not found: ID does not exist" containerID="200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7" Sep 29 21:56:11 crc kubenswrapper[4741]: I0929 21:56:11.596529 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7"} err="failed to get container status \"200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7\": rpc error: code = NotFound desc = could not find container \"200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7\": container with ID starting with 200a09b391f9f9da96f6e43413662c6cda5b4a76b115986e791073c1fbc54bb7 not found: ID does not exist" Sep 29 21:56:13 crc kubenswrapper[4741]: I0929 21:56:13.098183 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" path="/var/lib/kubelet/pods/c37c96d5-cca1-4f84-a834-55944b188eda/volumes" Sep 29 21:56:31 crc kubenswrapper[4741]: I0929 21:56:31.738681 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:56:31 crc kubenswrapper[4741]: I0929 21:56:31.739231 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.739242 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.740027 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.740081 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.741139 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdc380dcd36ba60be6b9a434fe90ce73ff08cd0c48d301bfb2e40232d9aeccef"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.741206 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://fdc380dcd36ba60be6b9a434fe90ce73ff08cd0c48d301bfb2e40232d9aeccef" gracePeriod=600 Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.994463 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="fdc380dcd36ba60be6b9a434fe90ce73ff08cd0c48d301bfb2e40232d9aeccef" exitCode=0 Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.994540 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"fdc380dcd36ba60be6b9a434fe90ce73ff08cd0c48d301bfb2e40232d9aeccef"} Sep 29 21:57:01 crc kubenswrapper[4741]: I0929 21:57:01.994779 4741 scope.go:117] "RemoveContainer" containerID="07f07717068b17476368e1e1090cb0680d1085b6291d1eae10a302c6b7ca3d90" Sep 29 21:57:03 crc kubenswrapper[4741]: I0929 21:57:03.006432 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a"} Sep 29 21:57:11 crc kubenswrapper[4741]: I0929 21:57:11.548052 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5178344b-4d12-4044-bf16-ea6eb56dc4f7/init-config-reloader/0.log" Sep 29 21:57:11 crc kubenswrapper[4741]: I0929 21:57:11.735969 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5178344b-4d12-4044-bf16-ea6eb56dc4f7/init-config-reloader/0.log" Sep 29 21:57:11 crc kubenswrapper[4741]: I0929 21:57:11.817483 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5178344b-4d12-4044-bf16-ea6eb56dc4f7/alertmanager/0.log" Sep 29 21:57:11 crc kubenswrapper[4741]: I0929 21:57:11.901977 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_5178344b-4d12-4044-bf16-ea6eb56dc4f7/config-reloader/0.log" Sep 29 21:57:12 crc kubenswrapper[4741]: I0929 21:57:12.072372 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7a4c3530-6b9e-47e4-a021-1abb149804b0/aodh-api/0.log" Sep 29 21:57:12 crc kubenswrapper[4741]: I0929 21:57:12.261132 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7a4c3530-6b9e-47e4-a021-1abb149804b0/aodh-listener/0.log" Sep 29 21:57:12 crc kubenswrapper[4741]: I0929 21:57:12.313197 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7a4c3530-6b9e-47e4-a021-1abb149804b0/aodh-evaluator/0.log" Sep 29 21:57:12 crc kubenswrapper[4741]: I0929 21:57:12.456234 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7a4c3530-6b9e-47e4-a021-1abb149804b0/aodh-notifier/0.log" Sep 29 21:57:12 crc kubenswrapper[4741]: I0929 21:57:12.664732 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-548cf4dc5b-fdx6w_83f5cab2-dd60-4453-88b6-8416abfc38e2/barbican-api/0.log" Sep 29 21:57:12 crc kubenswrapper[4741]: I0929 21:57:12.752575 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-548cf4dc5b-fdx6w_83f5cab2-dd60-4453-88b6-8416abfc38e2/barbican-api-log/0.log" Sep 29 21:57:12 crc kubenswrapper[4741]: I0929 21:57:12.909725 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-c97bb94fd-fj97f_c51fb5ea-8cca-41fe-bb13-5ffeea42eb99/barbican-keystone-listener/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.089725 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-c97bb94fd-fj97f_c51fb5ea-8cca-41fe-bb13-5ffeea42eb99/barbican-keystone-listener-log/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.284764 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69ddc8b4f7-ll8mc_804b72d0-4fb6-4e38-a52d-18703a62dbce/barbican-worker/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.345738 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69ddc8b4f7-ll8mc_804b72d0-4fb6-4e38-a52d-18703a62dbce/barbican-worker-log/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.559302 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-wfzrx_f4fb4b29-ab75-4ef4-be6d-b0c98cf28748/bootstrap-openstack-openstack-cell1/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.775318 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d5194444-589f-4132-8696-ccac62962087/ceilometer-central-agent/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.779245 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d5194444-589f-4132-8696-ccac62962087/ceilometer-notification-agent/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.924249 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d5194444-589f-4132-8696-ccac62962087/proxy-httpd/0.log" Sep 29 21:57:13 crc kubenswrapper[4741]: I0929 21:57:13.966414 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d5194444-589f-4132-8696-ccac62962087/sg-core/0.log" Sep 29 21:57:14 crc kubenswrapper[4741]: I0929 21:57:14.121185 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-n5fvd_6f58ba99-974b-4ace-bb5a-d01bb707de07/ceph-client-openstack-openstack-cell1/0.log" Sep 29 21:57:14 crc kubenswrapper[4741]: I0929 21:57:14.250545 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2bbfd908-1ced-4cae-8f3d-d7c89ec47c15/cinder-api/0.log" Sep 29 21:57:14 crc kubenswrapper[4741]: I0929 21:57:14.340661 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2bbfd908-1ced-4cae-8f3d-d7c89ec47c15/cinder-api-log/0.log" Sep 29 21:57:14 crc kubenswrapper[4741]: I0929 21:57:14.603005 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_df7971f7-6713-455c-acad-de1657b4d940/cinder-backup/0.log" Sep 29 21:57:14 crc kubenswrapper[4741]: I0929 21:57:14.603897 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_df7971f7-6713-455c-acad-de1657b4d940/probe/0.log" Sep 29 21:57:14 crc kubenswrapper[4741]: I0929 21:57:14.787934 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04/cinder-scheduler/0.log" Sep 29 21:57:14 crc kubenswrapper[4741]: I0929 21:57:14.805297 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ba97b0c5-5fa9-42b4-bb96-a02a5c8f8b04/probe/0.log" Sep 29 21:57:15 crc kubenswrapper[4741]: I0929 21:57:15.007848 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_4554f451-c9b8-44a6-a5cc-efa6837aa62a/cinder-volume/0.log" Sep 29 21:57:15 crc kubenswrapper[4741]: I0929 21:57:15.054804 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_4554f451-c9b8-44a6-a5cc-efa6837aa62a/probe/0.log" Sep 29 21:57:15 crc kubenswrapper[4741]: I0929 21:57:15.659592 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-799wv_437a18a3-85a4-4c95-87cd-bfc208f23a23/configure-network-openstack-openstack-cell1/0.log" Sep 29 21:57:15 crc kubenswrapper[4741]: I0929 21:57:15.864475 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-hwlcx_f372c5ae-6094-4c36-8b23-0a06a4e50c93/configure-os-openstack-openstack-cell1/0.log" Sep 29 21:57:15 crc kubenswrapper[4741]: I0929 21:57:15.972557 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5995b898cf-xwc7d_cb3a3160-8aa9-49f5-9479-208262519896/init/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.105991 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5995b898cf-xwc7d_cb3a3160-8aa9-49f5-9479-208262519896/init/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.178005 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5995b898cf-xwc7d_cb3a3160-8aa9-49f5-9479-208262519896/dnsmasq-dns/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.287206 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-g5757_80e293dd-446d-4869-babe-815944d08cdb/download-cache-openstack-openstack-cell1/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.370720 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c5bf2f12-6a59-4aee-84f5-62c964edab86/glance-httpd/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.475842 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c5bf2f12-6a59-4aee-84f5-62c964edab86/glance-log/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.595069 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f9554882-6011-402e-b598-e15f0284c296/glance-httpd/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.664164 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f9554882-6011-402e-b598-e15f0284c296/glance-log/0.log" Sep 29 21:57:16 crc kubenswrapper[4741]: I0929 21:57:16.914313 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-67f555958f-bpf7z_c8d784d6-ef0b-4f31-a1c6-cbc08cbd2407/heat-api/0.log" Sep 29 21:57:17 crc kubenswrapper[4741]: I0929 21:57:17.141052 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-6df6b5f465-mp74v_c94210e6-f88d-472f-9048-5511f940db0c/heat-cfnapi/0.log" Sep 29 21:57:17 crc kubenswrapper[4741]: I0929 21:57:17.149752 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7695685f8c-jtmfd_65abba08-dad1-4b41-a663-4d56c2b152e4/heat-engine/0.log" Sep 29 21:57:17 crc kubenswrapper[4741]: I0929 21:57:17.830674 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6f54844d7c-rrjpm_6195fc5e-a6c4-4a78-b45e-5c90ef096e3f/horizon-log/0.log" Sep 29 21:57:17 crc kubenswrapper[4741]: I0929 21:57:17.873397 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6f54844d7c-rrjpm_6195fc5e-a6c4-4a78-b45e-5c90ef096e3f/horizon/0.log" Sep 29 21:57:18 crc kubenswrapper[4741]: I0929 21:57:18.101453 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-tl5j6_5d41613b-ad82-41fa-893a-e150f9eccb58/install-certs-openstack-openstack-cell1/0.log" Sep 29 21:57:18 crc kubenswrapper[4741]: I0929 21:57:18.210050 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-krgtk_eac2989f-b47f-4ded-9a54-b8abccfb14bb/install-os-openstack-openstack-cell1/0.log" Sep 29 21:57:18 crc kubenswrapper[4741]: I0929 21:57:18.467903 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c569dd67b-5gb8r_6b9e24cb-db42-4298-ada9-940790341bcd/keystone-api/0.log" Sep 29 21:57:18 crc kubenswrapper[4741]: I0929 21:57:18.530503 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29319661-lblwq_51eef39d-57c6-4886-9b55-3378d8527072/keystone-cron/0.log" Sep 29 21:57:18 crc kubenswrapper[4741]: I0929 21:57:18.647545 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c998616e-1aed-4b70-a3fe-05d625649fa4/kube-state-metrics/0.log" Sep 29 21:57:18 crc kubenswrapper[4741]: I0929 21:57:18.877869 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-7qvfs_94f70bce-ae42-46eb-b72e-3c7d105608d3/libvirt-openstack-openstack-cell1/0.log" Sep 29 21:57:19 crc kubenswrapper[4741]: I0929 21:57:19.049132 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6/manila-api-log/0.log" Sep 29 21:57:19 crc kubenswrapper[4741]: I0929 21:57:19.135006 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_9b2c97de-7745-4f47-9ea9-fbe4bd1e8cb6/manila-api/0.log" Sep 29 21:57:19 crc kubenswrapper[4741]: I0929 21:57:19.339977 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_cf1826cd-d5f3-4ed1-ad4e-9646edde3a65/probe/0.log" Sep 29 21:57:19 crc kubenswrapper[4741]: I0929 21:57:19.366868 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_cf1826cd-d5f3-4ed1-ad4e-9646edde3a65/manila-scheduler/0.log" Sep 29 21:57:19 crc kubenswrapper[4741]: I0929 21:57:19.554117 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a28a5999-6348-4d83-9605-16292704fb39/probe/0.log" Sep 29 21:57:19 crc kubenswrapper[4741]: I0929 21:57:19.578211 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a28a5999-6348-4d83-9605-16292704fb39/manila-share/0.log" Sep 29 21:57:20 crc kubenswrapper[4741]: I0929 21:57:20.041631 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-ccc98d775-csfpc_1a4e588c-5749-4258-9986-b32c09909091/neutron-api/0.log" Sep 29 21:57:20 crc kubenswrapper[4741]: I0929 21:57:20.073081 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-ccc98d775-csfpc_1a4e588c-5749-4258-9986-b32c09909091/neutron-httpd/0.log" Sep 29 21:57:20 crc kubenswrapper[4741]: I0929 21:57:20.379557 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-s6xrg_6b5b0fa0-ac2f-4984-9585-304225817db7/neutron-dhcp-openstack-openstack-cell1/0.log" Sep 29 21:57:20 crc kubenswrapper[4741]: I0929 21:57:20.664708 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-27c4v_50889d55-3f51-469f-9dfb-e08d0fef628d/neutron-metadata-openstack-openstack-cell1/0.log" Sep 29 21:57:20 crc kubenswrapper[4741]: I0929 21:57:20.950920 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-qdcpx_19605586-361e-4c9a-9d60-685fd74ddb71/neutron-sriov-openstack-openstack-cell1/0.log" Sep 29 21:57:21 crc kubenswrapper[4741]: I0929 21:57:21.210625 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0863919b-b3fc-40b0-9957-8cc9573c2e3b/nova-api-api/0.log" Sep 29 21:57:21 crc kubenswrapper[4741]: I0929 21:57:21.333629 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0863919b-b3fc-40b0-9957-8cc9573c2e3b/nova-api-log/0.log" Sep 29 21:57:21 crc kubenswrapper[4741]: I0929 21:57:21.595909 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0bc6cc49-e74f-455c-9a7b-090cba934347/nova-cell0-conductor-conductor/0.log" Sep 29 21:57:21 crc kubenswrapper[4741]: I0929 21:57:21.907102 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ce05df11-1014-49e8-8744-3adcb02eb50b/nova-cell1-conductor-conductor/0.log" Sep 29 21:57:22 crc kubenswrapper[4741]: I0929 21:57:22.268401 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3454c3e1-18f9-425f-a9fa-7403c89a128b/nova-cell1-novncproxy-novncproxy/0.log" Sep 29 21:57:22 crc kubenswrapper[4741]: I0929 21:57:22.626773 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkcrt8_b6c8b549-a934-4c9d-919c-b5dee30d4385/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Sep 29 21:57:22 crc kubenswrapper[4741]: I0929 21:57:22.720986 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b1e2f180-603f-4760-8506-eb10f26199fe/memcached/0.log" Sep 29 21:57:22 crc kubenswrapper[4741]: I0929 21:57:22.987660 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-qqgg6_2f4bee73-3b5c-46f4-9ab7-1631dc495df3/nova-cell1-openstack-openstack-cell1/0.log" Sep 29 21:57:22 crc kubenswrapper[4741]: I0929 21:57:22.989111 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_aa65212c-f217-460f-adda-71813980fb4c/nova-metadata-log/0.log" Sep 29 21:57:23 crc kubenswrapper[4741]: I0929 21:57:23.167159 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_aa65212c-f217-460f-adda-71813980fb4c/nova-metadata-metadata/0.log" Sep 29 21:57:23 crc kubenswrapper[4741]: I0929 21:57:23.238319 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_001b7218-494b-4004-ad8c-4474aabad9dd/nova-scheduler-scheduler/0.log" Sep 29 21:57:23 crc kubenswrapper[4741]: I0929 21:57:23.575333 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-868985f9c4-bh7cz_5d39261b-e2bb-4ff2-8e37-1e91d0161c3e/init/0.log" Sep 29 21:57:23 crc kubenswrapper[4741]: I0929 21:57:23.743984 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-868985f9c4-bh7cz_5d39261b-e2bb-4ff2-8e37-1e91d0161c3e/init/0.log" Sep 29 21:57:23 crc kubenswrapper[4741]: I0929 21:57:23.788659 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-868985f9c4-bh7cz_5d39261b-e2bb-4ff2-8e37-1e91d0161c3e/octavia-api-provider-agent/0.log" Sep 29 21:57:23 crc kubenswrapper[4741]: I0929 21:57:23.962816 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-868985f9c4-bh7cz_5d39261b-e2bb-4ff2-8e37-1e91d0161c3e/octavia-api/0.log" Sep 29 21:57:23 crc kubenswrapper[4741]: I0929 21:57:23.994900 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-nsg8t_13b51e85-b858-4c7d-b92f-d082220fae38/init/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.151371 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-nsg8t_13b51e85-b858-4c7d-b92f-d082220fae38/init/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.222837 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-nsg8t_13b51e85-b858-4c7d-b92f-d082220fae38/octavia-healthmanager/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.392261 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-v24jm_dfce8c59-065c-4d36-a85d-ddbb32dc6bf9/init/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.534491 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-v24jm_dfce8c59-065c-4d36-a85d-ddbb32dc6bf9/octavia-housekeeping/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.549249 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-v24jm_dfce8c59-065c-4d36-a85d-ddbb32dc6bf9/init/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.717454 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-vmzlp_412eba2a-c36e-42a4-a15f-661fbf0d902c/init/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.861662 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-vmzlp_412eba2a-c36e-42a4-a15f-661fbf0d902c/octavia-amphora-httpd/0.log" Sep 29 21:57:24 crc kubenswrapper[4741]: I0929 21:57:24.897665 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-vmzlp_412eba2a-c36e-42a4-a15f-661fbf0d902c/init/0.log" Sep 29 21:57:25 crc kubenswrapper[4741]: I0929 21:57:25.037381 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-gjbhp_1f0963e9-4b39-4a0a-8ba5-7f47226efd00/init/0.log" Sep 29 21:57:25 crc kubenswrapper[4741]: I0929 21:57:25.661178 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-gjbhp_1f0963e9-4b39-4a0a-8ba5-7f47226efd00/init/0.log" Sep 29 21:57:25 crc kubenswrapper[4741]: I0929 21:57:25.696357 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-gjbhp_1f0963e9-4b39-4a0a-8ba5-7f47226efd00/octavia-rsyslog/0.log" Sep 29 21:57:25 crc kubenswrapper[4741]: I0929 21:57:25.831995 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-744s7_fba4561b-c4ff-4b6b-8785-e7605776aaf2/init/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.046552 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-744s7_fba4561b-c4ff-4b6b-8785-e7605776aaf2/init/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.104821 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-744s7_fba4561b-c4ff-4b6b-8785-e7605776aaf2/octavia-worker/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.235934 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f79ca3c3-1af6-4776-bc53-fffef15a23d9/mysql-bootstrap/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.391236 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f79ca3c3-1af6-4776-bc53-fffef15a23d9/mysql-bootstrap/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.401426 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f79ca3c3-1af6-4776-bc53-fffef15a23d9/galera/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.559532 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d2b041ad-6a3d-43f2-8faa-392b7c05c539/mysql-bootstrap/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.753116 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d2b041ad-6a3d-43f2-8faa-392b7c05c539/galera/0.log" Sep 29 21:57:26 crc kubenswrapper[4741]: I0929 21:57:26.757826 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d2b041ad-6a3d-43f2-8faa-392b7c05c539/mysql-bootstrap/0.log" Sep 29 21:57:27 crc kubenswrapper[4741]: I0929 21:57:27.551997 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gl88h_9cc73ea6-7c30-4a29-be40-7b85ceaeff71/openstack-network-exporter/0.log" Sep 29 21:57:27 crc kubenswrapper[4741]: I0929 21:57:27.585444 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_786fbbc4-3a41-4732-8ec1-c948fa4a346f/openstackclient/0.log" Sep 29 21:57:27 crc kubenswrapper[4741]: I0929 21:57:27.782593 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mlwzz_32d7ebdb-b4ae-4d70-ace7-458c7334b352/ovsdb-server-init/0.log" Sep 29 21:57:27 crc kubenswrapper[4741]: I0929 21:57:27.938146 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mlwzz_32d7ebdb-b4ae-4d70-ace7-458c7334b352/ovsdb-server-init/0.log" Sep 29 21:57:27 crc kubenswrapper[4741]: I0929 21:57:27.953011 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mlwzz_32d7ebdb-b4ae-4d70-ace7-458c7334b352/ovs-vswitchd/0.log" Sep 29 21:57:27 crc kubenswrapper[4741]: I0929 21:57:27.965596 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mlwzz_32d7ebdb-b4ae-4d70-ace7-458c7334b352/ovsdb-server/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.124142 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xbqvm_8755cb82-1296-4625-8767-4033013502ca/ovn-controller/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.316932 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4ca6fc6d-897d-4994-abc7-890d66323a76/openstack-network-exporter/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.361236 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4ca6fc6d-897d-4994-abc7-890d66323a76/ovn-northd/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.567688 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-8lmzv_3b400c3e-1f0f-4283-8e0f-b62978f3877d/ovn-openstack-openstack-cell1/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.709315 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2e48c5ed-877d-43d5-8a15-f47472801089/openstack-network-exporter/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.773774 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2e48c5ed-877d-43d5-8a15-f47472801089/ovsdbserver-nb/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.893329 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_77eb4b78-487a-4078-91b3-98172b274a9f/openstack-network-exporter/0.log" Sep 29 21:57:28 crc kubenswrapper[4741]: I0929 21:57:28.982695 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_77eb4b78-487a-4078-91b3-98172b274a9f/ovsdbserver-nb/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.062952 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_a7bddc92-7548-4ed5-9aa1-703d220dee0f/openstack-network-exporter/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.156948 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_a7bddc92-7548-4ed5-9aa1-703d220dee0f/ovsdbserver-nb/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.289504 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c00b2f37-4126-49e1-99e6-4cf54d225de1/openstack-network-exporter/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.348736 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c00b2f37-4126-49e1-99e6-4cf54d225de1/ovsdbserver-sb/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.480505 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e5583f47-e04c-4bc3-8614-4c2d67f61a8b/openstack-network-exporter/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.522842 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e5583f47-e04c-4bc3-8614-4c2d67f61a8b/ovsdbserver-sb/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.686305 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_ffa3c586-6f20-41bd-a638-69c1511d7652/openstack-network-exporter/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.693271 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_ffa3c586-6f20-41bd-a638-69c1511d7652/ovsdbserver-sb/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.948576 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-649f4d5b76-lcsk7_3233508f-ad82-4a1e-aa95-d647a3fdff0b/placement-log/0.log" Sep 29 21:57:29 crc kubenswrapper[4741]: I0929 21:57:29.964352 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-649f4d5b76-lcsk7_3233508f-ad82-4a1e-aa95-d647a3fdff0b/placement-api/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.152917 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-ccwp85_3e541050-007b-407c-9a45-2eb70ade11d9/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.344145 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8396f12c-5cd2-4db4-b2cb-0fa25e63efaf/init-config-reloader/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.496978 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8396f12c-5cd2-4db4-b2cb-0fa25e63efaf/config-reloader/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.532497 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8396f12c-5cd2-4db4-b2cb-0fa25e63efaf/prometheus/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.537604 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8396f12c-5cd2-4db4-b2cb-0fa25e63efaf/init-config-reloader/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.669175 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8396f12c-5cd2-4db4-b2cb-0fa25e63efaf/thanos-sidecar/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.732358 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844/setup-container/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.916482 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844/setup-container/0.log" Sep 29 21:57:30 crc kubenswrapper[4741]: I0929 21:57:30.929803 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b5e09dd9-bbbc-4efa-a28c-2d6aaf4fe844/rabbitmq/0.log" Sep 29 21:57:31 crc kubenswrapper[4741]: I0929 21:57:31.086310 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7b838fcc-8842-45a4-bbfe-29db7765bda8/setup-container/0.log" Sep 29 21:57:31 crc kubenswrapper[4741]: I0929 21:57:31.233329 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7b838fcc-8842-45a4-bbfe-29db7765bda8/setup-container/0.log" Sep 29 21:57:31 crc kubenswrapper[4741]: I0929 21:57:31.266823 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7b838fcc-8842-45a4-bbfe-29db7765bda8/rabbitmq/0.log" Sep 29 21:57:31 crc kubenswrapper[4741]: I0929 21:57:31.414244 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-562bz_fb553e7a-d9f7-4eb9-8aba-8452d128727c/reboot-os-openstack-openstack-cell1/0.log" Sep 29 21:57:31 crc kubenswrapper[4741]: I0929 21:57:31.521679 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-7jxgv_7dfada48-0466-4d17-9ef0-a6ffc08f1035/run-os-openstack-openstack-cell1/0.log" Sep 29 21:57:31 crc kubenswrapper[4741]: I0929 21:57:31.711523 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-c9bsh_b9b60bc5-0ab8-46c2-adeb-158cc32b9982/ssh-known-hosts-openstack/0.log" Sep 29 21:57:31 crc kubenswrapper[4741]: I0929 21:57:31.867679 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-28fg5_02f07af8-50b9-4252-90b9-a04657a5916d/telemetry-openstack-openstack-cell1/0.log" Sep 29 21:57:32 crc kubenswrapper[4741]: I0929 21:57:32.024867 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-jzhc5_e3d3a768-7bc7-4c8c-ade1-f8ddcabd8261/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Sep 29 21:57:32 crc kubenswrapper[4741]: I0929 21:57:32.182367 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-wqv85_b531de97-929c-4a21-af0b-fa956d7b6cc6/validate-network-openstack-openstack-cell1/0.log" Sep 29 21:58:08 crc kubenswrapper[4741]: I0929 21:58:08.733028 4741 generic.go:334] "Generic (PLEG): container finished" podID="1cd3776d-f1db-4f09-a62e-3ffc7a681187" containerID="d3e60fb8e9ffa9895a8f9f9932b18b3fd6ad6c0b655dcb0ddcac2c63545f8b8d" exitCode=0 Sep 29 21:58:08 crc kubenswrapper[4741]: I0929 21:58:08.733559 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-pqhls" event={"ID":"1cd3776d-f1db-4f09-a62e-3ffc7a681187","Type":"ContainerDied","Data":"d3e60fb8e9ffa9895a8f9f9932b18b3fd6ad6c0b655dcb0ddcac2c63545f8b8d"} Sep 29 21:58:09 crc kubenswrapper[4741]: I0929 21:58:09.873099 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:58:09 crc kubenswrapper[4741]: I0929 21:58:09.948461 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4m44r/crc-debug-pqhls"] Sep 29 21:58:09 crc kubenswrapper[4741]: I0929 21:58:09.960104 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4m44r/crc-debug-pqhls"] Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.025041 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cd3776d-f1db-4f09-a62e-3ffc7a681187-host\") pod \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.025421 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpg2x\" (UniqueName: \"kubernetes.io/projected/1cd3776d-f1db-4f09-a62e-3ffc7a681187-kube-api-access-jpg2x\") pod \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\" (UID: \"1cd3776d-f1db-4f09-a62e-3ffc7a681187\") " Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.025537 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1cd3776d-f1db-4f09-a62e-3ffc7a681187-host" (OuterVolumeSpecName: "host") pod "1cd3776d-f1db-4f09-a62e-3ffc7a681187" (UID: "1cd3776d-f1db-4f09-a62e-3ffc7a681187"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.025843 4741 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cd3776d-f1db-4f09-a62e-3ffc7a681187-host\") on node \"crc\" DevicePath \"\"" Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.057225 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cd3776d-f1db-4f09-a62e-3ffc7a681187-kube-api-access-jpg2x" (OuterVolumeSpecName: "kube-api-access-jpg2x") pod "1cd3776d-f1db-4f09-a62e-3ffc7a681187" (UID: "1cd3776d-f1db-4f09-a62e-3ffc7a681187"). InnerVolumeSpecName "kube-api-access-jpg2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.127452 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpg2x\" (UniqueName: \"kubernetes.io/projected/1cd3776d-f1db-4f09-a62e-3ffc7a681187-kube-api-access-jpg2x\") on node \"crc\" DevicePath \"\"" Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.771772 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="696cf9dbc44e1097d3c0ad5c48f1ef738139e79471b78962d0e1f52ef13ad6ac" Sep 29 21:58:10 crc kubenswrapper[4741]: I0929 21:58:10.771950 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-pqhls" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.096216 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cd3776d-f1db-4f09-a62e-3ffc7a681187" path="/var/lib/kubelet/pods/1cd3776d-f1db-4f09-a62e-3ffc7a681187/volumes" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.329305 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4m44r/crc-debug-2npjm"] Sep 29 21:58:11 crc kubenswrapper[4741]: E0929 21:58:11.330284 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="extract-utilities" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.330351 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="extract-utilities" Sep 29 21:58:11 crc kubenswrapper[4741]: E0929 21:58:11.330494 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd3776d-f1db-4f09-a62e-3ffc7a681187" containerName="container-00" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.330556 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd3776d-f1db-4f09-a62e-3ffc7a681187" containerName="container-00" Sep 29 21:58:11 crc kubenswrapper[4741]: E0929 21:58:11.330655 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="extract-content" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.330714 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="extract-content" Sep 29 21:58:11 crc kubenswrapper[4741]: E0929 21:58:11.330779 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="registry-server" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.330830 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="registry-server" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.331126 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd3776d-f1db-4f09-a62e-3ffc7a681187" containerName="container-00" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.331209 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="c37c96d5-cca1-4f84-a834-55944b188eda" containerName="registry-server" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.332052 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.454124 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4stvj\" (UniqueName: \"kubernetes.io/projected/d15da0b4-dfbb-4f65-a99d-a639f50449d7-kube-api-access-4stvj\") pod \"crc-debug-2npjm\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.454410 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d15da0b4-dfbb-4f65-a99d-a639f50449d7-host\") pod \"crc-debug-2npjm\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.556183 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4stvj\" (UniqueName: \"kubernetes.io/projected/d15da0b4-dfbb-4f65-a99d-a639f50449d7-kube-api-access-4stvj\") pod \"crc-debug-2npjm\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.556349 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d15da0b4-dfbb-4f65-a99d-a639f50449d7-host\") pod \"crc-debug-2npjm\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.556515 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d15da0b4-dfbb-4f65-a99d-a639f50449d7-host\") pod \"crc-debug-2npjm\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.577576 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4stvj\" (UniqueName: \"kubernetes.io/projected/d15da0b4-dfbb-4f65-a99d-a639f50449d7-kube-api-access-4stvj\") pod \"crc-debug-2npjm\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.652153 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:11 crc kubenswrapper[4741]: I0929 21:58:11.783897 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-2npjm" event={"ID":"d15da0b4-dfbb-4f65-a99d-a639f50449d7","Type":"ContainerStarted","Data":"d5692e3dfa2066fd07a32025b1fdcc00ba44205166470767cfb7e4cbd8589ccf"} Sep 29 21:58:12 crc kubenswrapper[4741]: I0929 21:58:12.792699 4741 generic.go:334] "Generic (PLEG): container finished" podID="d15da0b4-dfbb-4f65-a99d-a639f50449d7" containerID="81a0debb7e1393db51cbfb2547eebe48f3b9601d667609e7ec7fad56ae3798d0" exitCode=0 Sep 29 21:58:12 crc kubenswrapper[4741]: I0929 21:58:12.792803 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-2npjm" event={"ID":"d15da0b4-dfbb-4f65-a99d-a639f50449d7","Type":"ContainerDied","Data":"81a0debb7e1393db51cbfb2547eebe48f3b9601d667609e7ec7fad56ae3798d0"} Sep 29 21:58:13 crc kubenswrapper[4741]: I0929 21:58:13.913895 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:13 crc kubenswrapper[4741]: I0929 21:58:13.999584 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4stvj\" (UniqueName: \"kubernetes.io/projected/d15da0b4-dfbb-4f65-a99d-a639f50449d7-kube-api-access-4stvj\") pod \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " Sep 29 21:58:13 crc kubenswrapper[4741]: I0929 21:58:13.999660 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d15da0b4-dfbb-4f65-a99d-a639f50449d7-host\") pod \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\" (UID: \"d15da0b4-dfbb-4f65-a99d-a639f50449d7\") " Sep 29 21:58:14 crc kubenswrapper[4741]: I0929 21:58:13.999816 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d15da0b4-dfbb-4f65-a99d-a639f50449d7-host" (OuterVolumeSpecName: "host") pod "d15da0b4-dfbb-4f65-a99d-a639f50449d7" (UID: "d15da0b4-dfbb-4f65-a99d-a639f50449d7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 21:58:14 crc kubenswrapper[4741]: I0929 21:58:14.000055 4741 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d15da0b4-dfbb-4f65-a99d-a639f50449d7-host\") on node \"crc\" DevicePath \"\"" Sep 29 21:58:14 crc kubenswrapper[4741]: I0929 21:58:14.008989 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15da0b4-dfbb-4f65-a99d-a639f50449d7-kube-api-access-4stvj" (OuterVolumeSpecName: "kube-api-access-4stvj") pod "d15da0b4-dfbb-4f65-a99d-a639f50449d7" (UID: "d15da0b4-dfbb-4f65-a99d-a639f50449d7"). InnerVolumeSpecName "kube-api-access-4stvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:58:14 crc kubenswrapper[4741]: I0929 21:58:14.101355 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4stvj\" (UniqueName: \"kubernetes.io/projected/d15da0b4-dfbb-4f65-a99d-a639f50449d7-kube-api-access-4stvj\") on node \"crc\" DevicePath \"\"" Sep 29 21:58:14 crc kubenswrapper[4741]: I0929 21:58:14.819161 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-2npjm" event={"ID":"d15da0b4-dfbb-4f65-a99d-a639f50449d7","Type":"ContainerDied","Data":"d5692e3dfa2066fd07a32025b1fdcc00ba44205166470767cfb7e4cbd8589ccf"} Sep 29 21:58:14 crc kubenswrapper[4741]: I0929 21:58:14.819197 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5692e3dfa2066fd07a32025b1fdcc00ba44205166470767cfb7e4cbd8589ccf" Sep 29 21:58:14 crc kubenswrapper[4741]: I0929 21:58:14.819245 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-2npjm" Sep 29 21:58:23 crc kubenswrapper[4741]: I0929 21:58:23.314771 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4m44r/crc-debug-2npjm"] Sep 29 21:58:23 crc kubenswrapper[4741]: I0929 21:58:23.324230 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4m44r/crc-debug-2npjm"] Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.499955 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4m44r/crc-debug-w79mf"] Sep 29 21:58:24 crc kubenswrapper[4741]: E0929 21:58:24.500759 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15da0b4-dfbb-4f65-a99d-a639f50449d7" containerName="container-00" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.500774 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15da0b4-dfbb-4f65-a99d-a639f50449d7" containerName="container-00" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.501049 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15da0b4-dfbb-4f65-a99d-a639f50449d7" containerName="container-00" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.502033 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.605526 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c939f2f-2044-4845-960b-0d3829442cd3-host\") pod \"crc-debug-w79mf\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.605874 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkrgf\" (UniqueName: \"kubernetes.io/projected/4c939f2f-2044-4845-960b-0d3829442cd3-kube-api-access-tkrgf\") pod \"crc-debug-w79mf\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.707045 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkrgf\" (UniqueName: \"kubernetes.io/projected/4c939f2f-2044-4845-960b-0d3829442cd3-kube-api-access-tkrgf\") pod \"crc-debug-w79mf\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.707188 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c939f2f-2044-4845-960b-0d3829442cd3-host\") pod \"crc-debug-w79mf\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.707311 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c939f2f-2044-4845-960b-0d3829442cd3-host\") pod \"crc-debug-w79mf\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.727766 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkrgf\" (UniqueName: \"kubernetes.io/projected/4c939f2f-2044-4845-960b-0d3829442cd3-kube-api-access-tkrgf\") pod \"crc-debug-w79mf\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.828625 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:24 crc kubenswrapper[4741]: I0929 21:58:24.926812 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-w79mf" event={"ID":"4c939f2f-2044-4845-960b-0d3829442cd3","Type":"ContainerStarted","Data":"a3ff3f084e57cc63352bdd0d91de3ad563d15a1ab3b20030ab6ce8b399df8926"} Sep 29 21:58:25 crc kubenswrapper[4741]: I0929 21:58:25.097356 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15da0b4-dfbb-4f65-a99d-a639f50449d7" path="/var/lib/kubelet/pods/d15da0b4-dfbb-4f65-a99d-a639f50449d7/volumes" Sep 29 21:58:25 crc kubenswrapper[4741]: I0929 21:58:25.938343 4741 generic.go:334] "Generic (PLEG): container finished" podID="4c939f2f-2044-4845-960b-0d3829442cd3" containerID="fb0e1da9927e8db3cd2da0bb67777c9468d0388e5346e7ce9dc51f0419f8738d" exitCode=0 Sep 29 21:58:25 crc kubenswrapper[4741]: I0929 21:58:25.938403 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/crc-debug-w79mf" event={"ID":"4c939f2f-2044-4845-960b-0d3829442cd3","Type":"ContainerDied","Data":"fb0e1da9927e8db3cd2da0bb67777c9468d0388e5346e7ce9dc51f0419f8738d"} Sep 29 21:58:26 crc kubenswrapper[4741]: I0929 21:58:26.004995 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4m44r/crc-debug-w79mf"] Sep 29 21:58:26 crc kubenswrapper[4741]: I0929 21:58:26.013938 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4m44r/crc-debug-w79mf"] Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.066525 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.256311 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c939f2f-2044-4845-960b-0d3829442cd3-host\") pod \"4c939f2f-2044-4845-960b-0d3829442cd3\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.256481 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c939f2f-2044-4845-960b-0d3829442cd3-host" (OuterVolumeSpecName: "host") pod "4c939f2f-2044-4845-960b-0d3829442cd3" (UID: "4c939f2f-2044-4845-960b-0d3829442cd3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.256538 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkrgf\" (UniqueName: \"kubernetes.io/projected/4c939f2f-2044-4845-960b-0d3829442cd3-kube-api-access-tkrgf\") pod \"4c939f2f-2044-4845-960b-0d3829442cd3\" (UID: \"4c939f2f-2044-4845-960b-0d3829442cd3\") " Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.257244 4741 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c939f2f-2044-4845-960b-0d3829442cd3-host\") on node \"crc\" DevicePath \"\"" Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.262533 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c939f2f-2044-4845-960b-0d3829442cd3-kube-api-access-tkrgf" (OuterVolumeSpecName: "kube-api-access-tkrgf") pod "4c939f2f-2044-4845-960b-0d3829442cd3" (UID: "4c939f2f-2044-4845-960b-0d3829442cd3"). InnerVolumeSpecName "kube-api-access-tkrgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.359369 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkrgf\" (UniqueName: \"kubernetes.io/projected/4c939f2f-2044-4845-960b-0d3829442cd3-kube-api-access-tkrgf\") on node \"crc\" DevicePath \"\"" Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.960770 4741 scope.go:117] "RemoveContainer" containerID="fb0e1da9927e8db3cd2da0bb67777c9468d0388e5346e7ce9dc51f0419f8738d" Sep 29 21:58:27 crc kubenswrapper[4741]: I0929 21:58:27.960897 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/crc-debug-w79mf" Sep 29 21:58:29 crc kubenswrapper[4741]: I0929 21:58:29.099866 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c939f2f-2044-4845-960b-0d3829442cd3" path="/var/lib/kubelet/pods/4c939f2f-2044-4845-960b-0d3829442cd3/volumes" Sep 29 21:59:12 crc kubenswrapper[4741]: I0929 21:59:12.775660 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27_b4f1fd1d-9a47-476a-8fda-ae251041b7dd/util/0.log" Sep 29 21:59:12 crc kubenswrapper[4741]: I0929 21:59:12.987568 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27_b4f1fd1d-9a47-476a-8fda-ae251041b7dd/pull/0.log" Sep 29 21:59:12 crc kubenswrapper[4741]: I0929 21:59:12.990955 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27_b4f1fd1d-9a47-476a-8fda-ae251041b7dd/pull/0.log" Sep 29 21:59:12 crc kubenswrapper[4741]: I0929 21:59:12.993152 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27_b4f1fd1d-9a47-476a-8fda-ae251041b7dd/util/0.log" Sep 29 21:59:13 crc kubenswrapper[4741]: I0929 21:59:13.209181 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27_b4f1fd1d-9a47-476a-8fda-ae251041b7dd/util/0.log" Sep 29 21:59:13 crc kubenswrapper[4741]: I0929 21:59:13.253338 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27_b4f1fd1d-9a47-476a-8fda-ae251041b7dd/extract/0.log" Sep 29 21:59:13 crc kubenswrapper[4741]: I0929 21:59:13.259271 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36f479244198d369216604e24c34202fed1733124994d1bce9be2cc53dw5l27_b4f1fd1d-9a47-476a-8fda-ae251041b7dd/pull/0.log" Sep 29 21:59:13 crc kubenswrapper[4741]: I0929 21:59:13.738179 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-ntpxd_9f3daf24-fa9f-48f9-9f0a-df998fa76551/manager/0.log" Sep 29 21:59:13 crc kubenswrapper[4741]: I0929 21:59:13.777817 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-ntpxd_9f3daf24-fa9f-48f9-9f0a-df998fa76551/kube-rbac-proxy/0.log" Sep 29 21:59:13 crc kubenswrapper[4741]: I0929 21:59:13.791568 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-dkh26_5637f538-f88c-4990-8e83-b40e3217c76d/kube-rbac-proxy/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.071775 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-dkh26_5637f538-f88c-4990-8e83-b40e3217c76d/manager/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.093195 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-6d7tn_00c10904-1ec2-4578-891e-2baa6ae5ceb7/kube-rbac-proxy/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.094378 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-6d7tn_00c10904-1ec2-4578-891e-2baa6ae5ceb7/manager/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.343709 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-vnpfp_2774171b-e907-4eee-a0cd-4f8c1871ad45/kube-rbac-proxy/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.532992 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-vnpfp_2774171b-e907-4eee-a0cd-4f8c1871ad45/manager/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.539086 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-t5t9r_d3f19145-bf8f-4bb9-81cd-3fa20f066144/kube-rbac-proxy/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.611520 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-t5t9r_d3f19145-bf8f-4bb9-81cd-3fa20f066144/manager/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.764278 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-zd6mn_a581380e-2a13-42a4-8a12-606c842cefbc/kube-rbac-proxy/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.807910 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-zd6mn_a581380e-2a13-42a4-8a12-606c842cefbc/manager/0.log" Sep 29 21:59:14 crc kubenswrapper[4741]: I0929 21:59:14.912796 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-p5rtq_e0b2da1a-98f1-4008-864a-ad40956aef50/kube-rbac-proxy/0.log" Sep 29 21:59:15 crc kubenswrapper[4741]: I0929 21:59:15.130853 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-f44nr_836c126e-917e-4b62-a2aa-7d4caac672e4/kube-rbac-proxy/0.log" Sep 29 21:59:15 crc kubenswrapper[4741]: I0929 21:59:15.149250 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-f44nr_836c126e-917e-4b62-a2aa-7d4caac672e4/manager/0.log" Sep 29 21:59:15 crc kubenswrapper[4741]: I0929 21:59:15.291113 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-p5rtq_e0b2da1a-98f1-4008-864a-ad40956aef50/manager/0.log" Sep 29 21:59:15 crc kubenswrapper[4741]: I0929 21:59:15.425334 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-4sjbr_641c74c0-7259-43c6-8946-4da1b24a0f56/kube-rbac-proxy/0.log" Sep 29 21:59:15 crc kubenswrapper[4741]: I0929 21:59:15.596658 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-4sjbr_641c74c0-7259-43c6-8946-4da1b24a0f56/manager/0.log" Sep 29 21:59:15 crc kubenswrapper[4741]: I0929 21:59:15.841910 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-q5jhx_eeafa38f-b03a-4f83-bed3-3799872a37af/manager/0.log" Sep 29 21:59:15 crc kubenswrapper[4741]: I0929 21:59:15.955527 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-q5jhx_eeafa38f-b03a-4f83-bed3-3799872a37af/kube-rbac-proxy/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.180095 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-sztfr_a1c0a727-272e-4ec0-be2f-b836a9c5d8d8/manager/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.208741 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-sztfr_a1c0a727-272e-4ec0-be2f-b836a9c5d8d8/kube-rbac-proxy/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.234024 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-tmr6s_886c1e0e-7708-4c0a-adb2-bd5fc9b90477/kube-rbac-proxy/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.332404 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-tmr6s_886c1e0e-7708-4c0a-adb2-bd5fc9b90477/manager/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.558656 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-25l7k_178cb996-2b77-4bb5-b9b4-58027e1d2974/kube-rbac-proxy/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.614056 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-qv4km_ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd/kube-rbac-proxy/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.656312 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-25l7k_178cb996-2b77-4bb5-b9b4-58027e1d2974/manager/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.795384 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-qv4km_ffbed5a0-a999-4e50-9071-3a5ef3d6e1bd/manager/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.907296 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-6jp4h_5627e7c3-9301-4b01-a0c6-81957c86f73d/kube-rbac-proxy/0.log" Sep 29 21:59:16 crc kubenswrapper[4741]: I0929 21:59:16.929864 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-6jp4h_5627e7c3-9301-4b01-a0c6-81957c86f73d/manager/0.log" Sep 29 21:59:17 crc kubenswrapper[4741]: I0929 21:59:17.169611 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-59f6b8f5c9-qwnsv_6f1b6ef8-4bf6-4016-8333-1afad380afc5/kube-rbac-proxy/0.log" Sep 29 21:59:17 crc kubenswrapper[4741]: I0929 21:59:17.227228 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-679cf76f6c-hvgtj_5ba00c60-cd9d-48ec-bbd2-e1202d32e161/kube-rbac-proxy/0.log" Sep 29 21:59:17 crc kubenswrapper[4741]: I0929 21:59:17.428376 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-679cf76f6c-hvgtj_5ba00c60-cd9d-48ec-bbd2-e1202d32e161/operator/0.log" Sep 29 21:59:17 crc kubenswrapper[4741]: I0929 21:59:17.506975 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-w7bqm_3a0612c3-8758-4102-80e3-8082cef12fea/registry-server/0.log" Sep 29 21:59:17 crc kubenswrapper[4741]: I0929 21:59:17.653919 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-2qvtw_3a82b226-cd76-4abc-9699-2f907e8d7927/kube-rbac-proxy/0.log" Sep 29 21:59:17 crc kubenswrapper[4741]: I0929 21:59:17.825595 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-ljgcq_552e8661-bf43-4966-9aaf-2a2f9e514a41/kube-rbac-proxy/0.log" Sep 29 21:59:17 crc kubenswrapper[4741]: I0929 21:59:17.878771 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-2qvtw_3a82b226-cd76-4abc-9699-2f907e8d7927/manager/0.log" Sep 29 21:59:18 crc kubenswrapper[4741]: I0929 21:59:18.020548 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-ljgcq_552e8661-bf43-4966-9aaf-2a2f9e514a41/manager/0.log" Sep 29 21:59:18 crc kubenswrapper[4741]: I0929 21:59:18.334371 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-ntkgq_fe55fc47-9587-4c83-ac3b-8623ef3fe9db/operator/0.log" Sep 29 21:59:18 crc kubenswrapper[4741]: I0929 21:59:18.545412 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-bds8v_3474d9f5-6b8b-4763-8271-1e024733ff25/kube-rbac-proxy/0.log" Sep 29 21:59:18 crc kubenswrapper[4741]: I0929 21:59:18.637952 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-bds8v_3474d9f5-6b8b-4763-8271-1e024733ff25/manager/0.log" Sep 29 21:59:18 crc kubenswrapper[4741]: I0929 21:59:18.707486 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-jrmzg_a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e/kube-rbac-proxy/0.log" Sep 29 21:59:18 crc kubenswrapper[4741]: I0929 21:59:18.926583 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-6l5q9_054b7215-807f-4ce0-bd9e-1cc564808ed6/manager/0.log" Sep 29 21:59:19 crc kubenswrapper[4741]: I0929 21:59:19.020172 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-6l5q9_054b7215-807f-4ce0-bd9e-1cc564808ed6/kube-rbac-proxy/0.log" Sep 29 21:59:19 crc kubenswrapper[4741]: I0929 21:59:19.105540 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-jrmzg_a29fcbc5-3ec5-42e2-98fe-7eb8d453a76e/manager/0.log" Sep 29 21:59:19 crc kubenswrapper[4741]: I0929 21:59:19.203498 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-hgwfs_ac5bfa1f-c963-47e3-aa13-0bca249dcc78/kube-rbac-proxy/0.log" Sep 29 21:59:19 crc kubenswrapper[4741]: I0929 21:59:19.279201 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-hgwfs_ac5bfa1f-c963-47e3-aa13-0bca249dcc78/manager/0.log" Sep 29 21:59:19 crc kubenswrapper[4741]: I0929 21:59:19.688327 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-59f6b8f5c9-qwnsv_6f1b6ef8-4bf6-4016-8333-1afad380afc5/manager/0.log" Sep 29 21:59:31 crc kubenswrapper[4741]: I0929 21:59:31.739526 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 21:59:31 crc kubenswrapper[4741]: I0929 21:59:31.740337 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 21:59:39 crc kubenswrapper[4741]: I0929 21:59:39.238915 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-b5zmm_12453f09-2b49-47af-9f73-8c18c93a1812/control-plane-machine-set-operator/0.log" Sep 29 21:59:39 crc kubenswrapper[4741]: I0929 21:59:39.484462 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nrp5d_a1ff7353-843d-41bf-86eb-96a315f31bc3/kube-rbac-proxy/0.log" Sep 29 21:59:39 crc kubenswrapper[4741]: I0929 21:59:39.516340 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nrp5d_a1ff7353-843d-41bf-86eb-96a315f31bc3/machine-api-operator/0.log" Sep 29 21:59:52 crc kubenswrapper[4741]: I0929 21:59:52.362322 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-b9vcv_29d3eaf7-2c6c-45d8-90c9-49c1f3ed05fd/cert-manager-controller/0.log" Sep 29 21:59:52 crc kubenswrapper[4741]: I0929 21:59:52.556608 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-g2rtf_6eca0f9c-731e-49ea-ae29-40125ae4c0e2/cert-manager-cainjector/0.log" Sep 29 21:59:52 crc kubenswrapper[4741]: I0929 21:59:52.605142 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-gfzk9_ddbc8c2b-e339-4850-aea0-094d3beaf857/cert-manager-webhook/0.log" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.167677 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj"] Sep 29 22:00:00 crc kubenswrapper[4741]: E0929 22:00:00.168759 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c939f2f-2044-4845-960b-0d3829442cd3" containerName="container-00" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.168775 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c939f2f-2044-4845-960b-0d3829442cd3" containerName="container-00" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.168982 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c939f2f-2044-4845-960b-0d3829442cd3" containerName="container-00" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.169767 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.173276 4741 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.173930 4741 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.184341 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj"] Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.221497 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prf45\" (UniqueName: \"kubernetes.io/projected/717e5e25-074e-466b-824a-00966cac623b-kube-api-access-prf45\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.221570 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717e5e25-074e-466b-824a-00966cac623b-config-volume\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.221663 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717e5e25-074e-466b-824a-00966cac623b-secret-volume\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.324435 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prf45\" (UniqueName: \"kubernetes.io/projected/717e5e25-074e-466b-824a-00966cac623b-kube-api-access-prf45\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.324511 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717e5e25-074e-466b-824a-00966cac623b-config-volume\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.324577 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717e5e25-074e-466b-824a-00966cac623b-secret-volume\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.325497 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717e5e25-074e-466b-824a-00966cac623b-config-volume\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.334740 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717e5e25-074e-466b-824a-00966cac623b-secret-volume\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.342098 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prf45\" (UniqueName: \"kubernetes.io/projected/717e5e25-074e-466b-824a-00966cac623b-kube-api-access-prf45\") pod \"collect-profiles-29319720-p87rj\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:00 crc kubenswrapper[4741]: I0929 22:00:00.503321 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:01 crc kubenswrapper[4741]: I0929 22:00:01.024471 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj"] Sep 29 22:00:01 crc kubenswrapper[4741]: W0929 22:00:01.032207 4741 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod717e5e25_074e_466b_824a_00966cac623b.slice/crio-25874e0c0f56b9c28f0c372bda9ee9e3dca7f0d337dbd81ac13e01cdba55daba WatchSource:0}: Error finding container 25874e0c0f56b9c28f0c372bda9ee9e3dca7f0d337dbd81ac13e01cdba55daba: Status 404 returned error can't find the container with id 25874e0c0f56b9c28f0c372bda9ee9e3dca7f0d337dbd81ac13e01cdba55daba Sep 29 22:00:01 crc kubenswrapper[4741]: I0929 22:00:01.738638 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 22:00:01 crc kubenswrapper[4741]: I0929 22:00:01.739115 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 22:00:01 crc kubenswrapper[4741]: I0929 22:00:01.870699 4741 generic.go:334] "Generic (PLEG): container finished" podID="717e5e25-074e-466b-824a-00966cac623b" containerID="4daf5bd9e06e119f404dda8dfbf951236963a7e87392e081d74275c84a5766b5" exitCode=0 Sep 29 22:00:01 crc kubenswrapper[4741]: I0929 22:00:01.870750 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" event={"ID":"717e5e25-074e-466b-824a-00966cac623b","Type":"ContainerDied","Data":"4daf5bd9e06e119f404dda8dfbf951236963a7e87392e081d74275c84a5766b5"} Sep 29 22:00:01 crc kubenswrapper[4741]: I0929 22:00:01.870783 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" event={"ID":"717e5e25-074e-466b-824a-00966cac623b","Type":"ContainerStarted","Data":"25874e0c0f56b9c28f0c372bda9ee9e3dca7f0d337dbd81ac13e01cdba55daba"} Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.263359 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.402750 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prf45\" (UniqueName: \"kubernetes.io/projected/717e5e25-074e-466b-824a-00966cac623b-kube-api-access-prf45\") pod \"717e5e25-074e-466b-824a-00966cac623b\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.403259 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717e5e25-074e-466b-824a-00966cac623b-secret-volume\") pod \"717e5e25-074e-466b-824a-00966cac623b\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.403361 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717e5e25-074e-466b-824a-00966cac623b-config-volume\") pod \"717e5e25-074e-466b-824a-00966cac623b\" (UID: \"717e5e25-074e-466b-824a-00966cac623b\") " Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.404154 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/717e5e25-074e-466b-824a-00966cac623b-config-volume" (OuterVolumeSpecName: "config-volume") pod "717e5e25-074e-466b-824a-00966cac623b" (UID: "717e5e25-074e-466b-824a-00966cac623b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.409026 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717e5e25-074e-466b-824a-00966cac623b-kube-api-access-prf45" (OuterVolumeSpecName: "kube-api-access-prf45") pod "717e5e25-074e-466b-824a-00966cac623b" (UID: "717e5e25-074e-466b-824a-00966cac623b"). InnerVolumeSpecName "kube-api-access-prf45". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.412148 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717e5e25-074e-466b-824a-00966cac623b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "717e5e25-074e-466b-824a-00966cac623b" (UID: "717e5e25-074e-466b-824a-00966cac623b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.506120 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prf45\" (UniqueName: \"kubernetes.io/projected/717e5e25-074e-466b-824a-00966cac623b-kube-api-access-prf45\") on node \"crc\" DevicePath \"\"" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.506157 4741 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717e5e25-074e-466b-824a-00966cac623b-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.506167 4741 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717e5e25-074e-466b-824a-00966cac623b-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.891093 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" event={"ID":"717e5e25-074e-466b-824a-00966cac623b","Type":"ContainerDied","Data":"25874e0c0f56b9c28f0c372bda9ee9e3dca7f0d337dbd81ac13e01cdba55daba"} Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.891134 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319720-p87rj" Sep 29 22:00:03 crc kubenswrapper[4741]: I0929 22:00:03.891141 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25874e0c0f56b9c28f0c372bda9ee9e3dca7f0d337dbd81ac13e01cdba55daba" Sep 29 22:00:04 crc kubenswrapper[4741]: I0929 22:00:04.350169 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl"] Sep 29 22:00:04 crc kubenswrapper[4741]: I0929 22:00:04.359835 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319675-xjtpl"] Sep 29 22:00:05 crc kubenswrapper[4741]: I0929 22:00:05.100850 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ab01c7-5aab-45fd-b45d-1e8508dd0c16" path="/var/lib/kubelet/pods/98ab01c7-5aab-45fd-b45d-1e8508dd0c16/volumes" Sep 29 22:00:05 crc kubenswrapper[4741]: I0929 22:00:05.200792 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-kcgmv_7111177c-0787-4b38-9858-82853f33807b/nmstate-console-plugin/0.log" Sep 29 22:00:05 crc kubenswrapper[4741]: I0929 22:00:05.400975 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-f8s2v_6fd3114f-fedf-4aa4-a230-2e00b6e7f6e5/nmstate-handler/0.log" Sep 29 22:00:05 crc kubenswrapper[4741]: I0929 22:00:05.434642 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-b6wcb_e4b27a34-8d13-4f3e-97ec-f1428f02aef8/kube-rbac-proxy/0.log" Sep 29 22:00:05 crc kubenswrapper[4741]: I0929 22:00:05.447128 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-b6wcb_e4b27a34-8d13-4f3e-97ec-f1428f02aef8/nmstate-metrics/0.log" Sep 29 22:00:05 crc kubenswrapper[4741]: I0929 22:00:05.623625 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-c96rz_6d63068c-acfa-46f8-a905-f21e7ff67e84/nmstate-webhook/0.log" Sep 29 22:00:05 crc kubenswrapper[4741]: I0929 22:00:05.697907 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-mfw46_6567df3f-ff1b-4627-9b01-52eadd11d93c/nmstate-operator/0.log" Sep 29 22:00:19 crc kubenswrapper[4741]: I0929 22:00:19.806609 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-lf5lv_b4d81ed4-fc0b-4f83-9207-2b6ab63ced26/kube-rbac-proxy/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.087183 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-frr-files/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.320314 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-frr-files/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.328291 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-reloader/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.353698 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-metrics/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.448983 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-lf5lv_b4d81ed4-fc0b-4f83-9207-2b6ab63ced26/controller/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.508614 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-reloader/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.741865 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-metrics/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.742554 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-frr-files/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.764270 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-reloader/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.770116 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-metrics/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.933797 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-reloader/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.978538 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-frr-files/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.981984 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/controller/0.log" Sep 29 22:00:20 crc kubenswrapper[4741]: I0929 22:00:20.983212 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/cp-metrics/0.log" Sep 29 22:00:21 crc kubenswrapper[4741]: I0929 22:00:21.163040 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/kube-rbac-proxy/0.log" Sep 29 22:00:21 crc kubenswrapper[4741]: I0929 22:00:21.173719 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/frr-metrics/0.log" Sep 29 22:00:21 crc kubenswrapper[4741]: I0929 22:00:21.235495 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/kube-rbac-proxy-frr/0.log" Sep 29 22:00:21 crc kubenswrapper[4741]: I0929 22:00:21.445969 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-hbchd_bb1b55e6-2bf2-43c0-9c5f-c5896a59bac5/frr-k8s-webhook-server/0.log" Sep 29 22:00:21 crc kubenswrapper[4741]: I0929 22:00:21.456850 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/reloader/0.log" Sep 29 22:00:21 crc kubenswrapper[4741]: I0929 22:00:21.704242 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-578897fd67-n2ssx_c8d5a915-940f-4fd2-9efd-7ee3a9dba397/manager/0.log" Sep 29 22:00:21 crc kubenswrapper[4741]: I0929 22:00:21.867754 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68f987f9f5-zpvf6_9a35855e-0119-4a1f-928a-51a39f9aa326/webhook-server/0.log" Sep 29 22:00:22 crc kubenswrapper[4741]: I0929 22:00:22.212323 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-njmrg_80a22421-8515-4f7f-8af1-8cd2904154a0/kube-rbac-proxy/0.log" Sep 29 22:00:23 crc kubenswrapper[4741]: I0929 22:00:23.148035 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-njmrg_80a22421-8515-4f7f-8af1-8cd2904154a0/speaker/0.log" Sep 29 22:00:24 crc kubenswrapper[4741]: I0929 22:00:24.597194 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cfqh_c1a026f2-db7d-4ff5-b79a-ce82c325f71c/frr/0.log" Sep 29 22:00:31 crc kubenswrapper[4741]: I0929 22:00:31.739191 4741 patch_prober.go:28] interesting pod/machine-config-daemon-rtxqk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 22:00:31 crc kubenswrapper[4741]: I0929 22:00:31.739919 4741 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 22:00:31 crc kubenswrapper[4741]: I0929 22:00:31.739970 4741 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" Sep 29 22:00:31 crc kubenswrapper[4741]: I0929 22:00:31.740822 4741 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a"} pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 22:00:31 crc kubenswrapper[4741]: I0929 22:00:31.740883 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerName="machine-config-daemon" containerID="cri-o://1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" gracePeriod=600 Sep 29 22:00:31 crc kubenswrapper[4741]: E0929 22:00:31.871246 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:00:32 crc kubenswrapper[4741]: I0929 22:00:32.239832 4741 generic.go:334] "Generic (PLEG): container finished" podID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" exitCode=0 Sep 29 22:00:32 crc kubenswrapper[4741]: I0929 22:00:32.239890 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerDied","Data":"1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a"} Sep 29 22:00:32 crc kubenswrapper[4741]: I0929 22:00:32.240727 4741 scope.go:117] "RemoveContainer" containerID="fdc380dcd36ba60be6b9a434fe90ce73ff08cd0c48d301bfb2e40232d9aeccef" Sep 29 22:00:32 crc kubenswrapper[4741]: I0929 22:00:32.241715 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:00:32 crc kubenswrapper[4741]: E0929 22:00:32.242114 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:00:34 crc kubenswrapper[4741]: I0929 22:00:34.001250 4741 scope.go:117] "RemoveContainer" containerID="694ffb09050eca501cdc4eed77bfc64b4b536ba062c1a04e40b53486cdbc74da" Sep 29 22:00:34 crc kubenswrapper[4741]: I0929 22:00:34.934012 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd_7c0e9a8f-564e-472b-9f9f-fb571beeade4/util/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.179267 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd_7c0e9a8f-564e-472b-9f9f-fb571beeade4/util/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.250649 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd_7c0e9a8f-564e-472b-9f9f-fb571beeade4/pull/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.265041 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd_7c0e9a8f-564e-472b-9f9f-fb571beeade4/pull/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.435516 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd_7c0e9a8f-564e-472b-9f9f-fb571beeade4/util/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.467709 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd_7c0e9a8f-564e-472b-9f9f-fb571beeade4/pull/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.475971 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zdsqd_7c0e9a8f-564e-472b-9f9f-fb571beeade4/extract/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.627806 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_12a902af-c903-4004-9e1a-f3b4ab1cae5f/util/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.821557 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_12a902af-c903-4004-9e1a-f3b4ab1cae5f/util/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.822517 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_12a902af-c903-4004-9e1a-f3b4ab1cae5f/pull/0.log" Sep 29 22:00:35 crc kubenswrapper[4741]: I0929 22:00:35.857171 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_12a902af-c903-4004-9e1a-f3b4ab1cae5f/pull/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.030103 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_12a902af-c903-4004-9e1a-f3b4ab1cae5f/util/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.040640 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_12a902af-c903-4004-9e1a-f3b4ab1cae5f/pull/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.068202 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc6lf9f_12a902af-c903-4004-9e1a-f3b4ab1cae5f/extract/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.221648 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5_ba55057a-3c2f-4ea0-b2c3-a9b1f896f039/util/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.401150 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5_ba55057a-3c2f-4ea0-b2c3-a9b1f896f039/pull/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.415528 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5_ba55057a-3c2f-4ea0-b2c3-a9b1f896f039/pull/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.445504 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5_ba55057a-3c2f-4ea0-b2c3-a9b1f896f039/util/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.597909 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5_ba55057a-3c2f-4ea0-b2c3-a9b1f896f039/pull/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.600127 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5_ba55057a-3c2f-4ea0-b2c3-a9b1f896f039/util/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.635846 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dhngf5_ba55057a-3c2f-4ea0-b2c3-a9b1f896f039/extract/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.797960 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bkw22_478c7368-6364-4c7f-89a3-9a6358c3318e/extract-utilities/0.log" Sep 29 22:00:36 crc kubenswrapper[4741]: I0929 22:00:36.988825 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bkw22_478c7368-6364-4c7f-89a3-9a6358c3318e/extract-content/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.003040 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bkw22_478c7368-6364-4c7f-89a3-9a6358c3318e/extract-utilities/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.043284 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bkw22_478c7368-6364-4c7f-89a3-9a6358c3318e/extract-content/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.254071 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bkw22_478c7368-6364-4c7f-89a3-9a6358c3318e/extract-utilities/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.294775 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bkw22_478c7368-6364-4c7f-89a3-9a6358c3318e/extract-content/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.545860 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-k4kw7_d135c17a-0ef7-4994-8125-019b0d6f4fde/extract-utilities/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.734467 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-k4kw7_d135c17a-0ef7-4994-8125-019b0d6f4fde/extract-utilities/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.744613 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-k4kw7_d135c17a-0ef7-4994-8125-019b0d6f4fde/extract-content/0.log" Sep 29 22:00:37 crc kubenswrapper[4741]: I0929 22:00:37.824581 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-k4kw7_d135c17a-0ef7-4994-8125-019b0d6f4fde/extract-content/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.038753 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-k4kw7_d135c17a-0ef7-4994-8125-019b0d6f4fde/extract-utilities/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.052999 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-k4kw7_d135c17a-0ef7-4994-8125-019b0d6f4fde/extract-content/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.277508 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv_d33f5fb0-1f2e-43a2-8b7a-0061df4f821d/util/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.412044 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv_d33f5fb0-1f2e-43a2-8b7a-0061df4f821d/util/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.420434 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv_d33f5fb0-1f2e-43a2-8b7a-0061df4f821d/pull/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.576438 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv_d33f5fb0-1f2e-43a2-8b7a-0061df4f821d/pull/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.704001 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv_d33f5fb0-1f2e-43a2-8b7a-0061df4f821d/util/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.760365 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv_d33f5fb0-1f2e-43a2-8b7a-0061df4f821d/pull/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.794674 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d968sdlv_d33f5fb0-1f2e-43a2-8b7a-0061df4f821d/extract/0.log" Sep 29 22:00:38 crc kubenswrapper[4741]: I0929 22:00:38.932038 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bkw22_478c7368-6364-4c7f-89a3-9a6358c3318e/registry-server/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.025661 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fhs5q_674ba05b-0612-46ca-b5b6-a91eb3950d27/marketplace-operator/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.168103 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ztxkt_df857807-c2a1-48da-9f68-ee9a9aafbb43/extract-utilities/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.402486 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ztxkt_df857807-c2a1-48da-9f68-ee9a9aafbb43/extract-content/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.409133 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ztxkt_df857807-c2a1-48da-9f68-ee9a9aafbb43/extract-content/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.432929 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ztxkt_df857807-c2a1-48da-9f68-ee9a9aafbb43/extract-utilities/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.660741 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ztxkt_df857807-c2a1-48da-9f68-ee9a9aafbb43/extract-utilities/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.766934 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ztxkt_df857807-c2a1-48da-9f68-ee9a9aafbb43/extract-content/0.log" Sep 29 22:00:39 crc kubenswrapper[4741]: I0929 22:00:39.896279 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bjmn7_074c0cb0-3256-405d-a1ec-9884ef100eff/extract-utilities/0.log" Sep 29 22:00:40 crc kubenswrapper[4741]: I0929 22:00:40.050896 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-k4kw7_d135c17a-0ef7-4994-8125-019b0d6f4fde/registry-server/0.log" Sep 29 22:00:40 crc kubenswrapper[4741]: I0929 22:00:40.135112 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bjmn7_074c0cb0-3256-405d-a1ec-9884ef100eff/extract-content/0.log" Sep 29 22:00:40 crc kubenswrapper[4741]: I0929 22:00:40.166522 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bjmn7_074c0cb0-3256-405d-a1ec-9884ef100eff/extract-utilities/0.log" Sep 29 22:00:40 crc kubenswrapper[4741]: I0929 22:00:40.192448 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bjmn7_074c0cb0-3256-405d-a1ec-9884ef100eff/extract-content/0.log" Sep 29 22:00:40 crc kubenswrapper[4741]: I0929 22:00:40.209174 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ztxkt_df857807-c2a1-48da-9f68-ee9a9aafbb43/registry-server/0.log" Sep 29 22:00:40 crc kubenswrapper[4741]: I0929 22:00:40.424130 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bjmn7_074c0cb0-3256-405d-a1ec-9884ef100eff/extract-content/0.log" Sep 29 22:00:40 crc kubenswrapper[4741]: I0929 22:00:40.438051 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bjmn7_074c0cb0-3256-405d-a1ec-9884ef100eff/extract-utilities/0.log" Sep 29 22:00:41 crc kubenswrapper[4741]: I0929 22:00:41.597590 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bjmn7_074c0cb0-3256-405d-a1ec-9884ef100eff/registry-server/0.log" Sep 29 22:00:44 crc kubenswrapper[4741]: I0929 22:00:44.086349 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:00:44 crc kubenswrapper[4741]: E0929 22:00:44.087080 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:00:51 crc kubenswrapper[4741]: I0929 22:00:51.834340 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-hzrg2_c6b5ed3e-0d65-49c3-b2af-5eec98429b23/prometheus-operator/0.log" Sep 29 22:00:52 crc kubenswrapper[4741]: I0929 22:00:52.000107 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6f89b4f898-65492_51c125d4-27fa-4b7e-a99c-2253fb9e3e33/prometheus-operator-admission-webhook/0.log" Sep 29 22:00:52 crc kubenswrapper[4741]: I0929 22:00:52.042934 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6f89b4f898-6zg7s_ce0ae04d-8ff8-4668-8319-b9a955b7a6ec/prometheus-operator-admission-webhook/0.log" Sep 29 22:00:52 crc kubenswrapper[4741]: I0929 22:00:52.226113 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-m97dn_a12fd15d-08fb-4df4-842a-513513640697/perses-operator/0.log" Sep 29 22:00:52 crc kubenswrapper[4741]: I0929 22:00:52.303372 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-8kkkd_a791644a-1784-4f93-9772-85672174eafa/operator/0.log" Sep 29 22:00:55 crc kubenswrapper[4741]: I0929 22:00:55.086203 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:00:55 crc kubenswrapper[4741]: E0929 22:00:55.086937 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.164149 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29319721-bhv2x"] Sep 29 22:01:00 crc kubenswrapper[4741]: E0929 22:01:00.165310 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717e5e25-074e-466b-824a-00966cac623b" containerName="collect-profiles" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.165323 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="717e5e25-074e-466b-824a-00966cac623b" containerName="collect-profiles" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.165558 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="717e5e25-074e-466b-824a-00966cac623b" containerName="collect-profiles" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.168446 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.186592 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29319721-bhv2x"] Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.238056 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xk4r\" (UniqueName: \"kubernetes.io/projected/b01432f5-24f4-45f9-81f9-32f05dea806b-kube-api-access-6xk4r\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.238127 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-fernet-keys\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.238162 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-combined-ca-bundle\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.238288 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-config-data\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.341420 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-config-data\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.341556 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xk4r\" (UniqueName: \"kubernetes.io/projected/b01432f5-24f4-45f9-81f9-32f05dea806b-kube-api-access-6xk4r\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.341591 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-fernet-keys\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.341626 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-combined-ca-bundle\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.349241 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-fernet-keys\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.349445 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-combined-ca-bundle\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.351624 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-config-data\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.365018 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xk4r\" (UniqueName: \"kubernetes.io/projected/b01432f5-24f4-45f9-81f9-32f05dea806b-kube-api-access-6xk4r\") pod \"keystone-cron-29319721-bhv2x\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:00 crc kubenswrapper[4741]: I0929 22:01:00.506193 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:01 crc kubenswrapper[4741]: I0929 22:01:01.054373 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29319721-bhv2x"] Sep 29 22:01:01 crc kubenswrapper[4741]: I0929 22:01:01.578695 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319721-bhv2x" event={"ID":"b01432f5-24f4-45f9-81f9-32f05dea806b","Type":"ContainerStarted","Data":"ed5e2b4cbe87025bf0adabbd8693482ae6801b731e2b59a49d3ab0694af8d0c9"} Sep 29 22:01:01 crc kubenswrapper[4741]: I0929 22:01:01.579233 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319721-bhv2x" event={"ID":"b01432f5-24f4-45f9-81f9-32f05dea806b","Type":"ContainerStarted","Data":"d6959883959accb547b0994b68da9361e94341cdb510b160e97575d50e7ed0e1"} Sep 29 22:01:01 crc kubenswrapper[4741]: I0929 22:01:01.609813 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29319721-bhv2x" podStartSLOduration=1.609798429 podStartE2EDuration="1.609798429s" podCreationTimestamp="2025-09-29 22:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 22:01:01.60666377 +0000 UTC m=+10303.254453102" watchObservedRunningTime="2025-09-29 22:01:01.609798429 +0000 UTC m=+10303.257587761" Sep 29 22:01:04 crc kubenswrapper[4741]: I0929 22:01:04.613192 4741 generic.go:334] "Generic (PLEG): container finished" podID="b01432f5-24f4-45f9-81f9-32f05dea806b" containerID="ed5e2b4cbe87025bf0adabbd8693482ae6801b731e2b59a49d3ab0694af8d0c9" exitCode=0 Sep 29 22:01:04 crc kubenswrapper[4741]: I0929 22:01:04.613755 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319721-bhv2x" event={"ID":"b01432f5-24f4-45f9-81f9-32f05dea806b","Type":"ContainerDied","Data":"ed5e2b4cbe87025bf0adabbd8693482ae6801b731e2b59a49d3ab0694af8d0c9"} Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.033453 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.192291 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-combined-ca-bundle\") pod \"b01432f5-24f4-45f9-81f9-32f05dea806b\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.192517 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-fernet-keys\") pod \"b01432f5-24f4-45f9-81f9-32f05dea806b\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.192630 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-config-data\") pod \"b01432f5-24f4-45f9-81f9-32f05dea806b\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.192778 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xk4r\" (UniqueName: \"kubernetes.io/projected/b01432f5-24f4-45f9-81f9-32f05dea806b-kube-api-access-6xk4r\") pod \"b01432f5-24f4-45f9-81f9-32f05dea806b\" (UID: \"b01432f5-24f4-45f9-81f9-32f05dea806b\") " Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.204691 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b01432f5-24f4-45f9-81f9-32f05dea806b" (UID: "b01432f5-24f4-45f9-81f9-32f05dea806b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.204698 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b01432f5-24f4-45f9-81f9-32f05dea806b-kube-api-access-6xk4r" (OuterVolumeSpecName: "kube-api-access-6xk4r") pod "b01432f5-24f4-45f9-81f9-32f05dea806b" (UID: "b01432f5-24f4-45f9-81f9-32f05dea806b"). InnerVolumeSpecName "kube-api-access-6xk4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.229307 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b01432f5-24f4-45f9-81f9-32f05dea806b" (UID: "b01432f5-24f4-45f9-81f9-32f05dea806b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.253707 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-config-data" (OuterVolumeSpecName: "config-data") pod "b01432f5-24f4-45f9-81f9-32f05dea806b" (UID: "b01432f5-24f4-45f9-81f9-32f05dea806b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.296795 4741 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.296827 4741 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.296840 4741 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b01432f5-24f4-45f9-81f9-32f05dea806b-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.296850 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xk4r\" (UniqueName: \"kubernetes.io/projected/b01432f5-24f4-45f9-81f9-32f05dea806b-kube-api-access-6xk4r\") on node \"crc\" DevicePath \"\"" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.690014 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319721-bhv2x" event={"ID":"b01432f5-24f4-45f9-81f9-32f05dea806b","Type":"ContainerDied","Data":"d6959883959accb547b0994b68da9361e94341cdb510b160e97575d50e7ed0e1"} Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.690490 4741 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6959883959accb547b0994b68da9361e94341cdb510b160e97575d50e7ed0e1" Sep 29 22:01:06 crc kubenswrapper[4741]: I0929 22:01:06.690279 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319721-bhv2x" Sep 29 22:01:06 crc kubenswrapper[4741]: E0929 22:01:06.933002 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb01432f5_24f4_45f9_81f9_32f05dea806b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb01432f5_24f4_45f9_81f9_32f05dea806b.slice/crio-d6959883959accb547b0994b68da9361e94341cdb510b160e97575d50e7ed0e1\": RecentStats: unable to find data in memory cache]" Sep 29 22:01:08 crc kubenswrapper[4741]: I0929 22:01:08.086474 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:01:08 crc kubenswrapper[4741]: E0929 22:01:08.087182 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:01:23 crc kubenswrapper[4741]: I0929 22:01:23.085881 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:01:23 crc kubenswrapper[4741]: E0929 22:01:23.086861 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:01:24 crc kubenswrapper[4741]: E0929 22:01:24.541816 4741 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.227:45882->38.102.83.227:37241: read tcp 38.102.83.227:45882->38.102.83.227:37241: read: connection reset by peer Sep 29 22:01:36 crc kubenswrapper[4741]: I0929 22:01:36.086813 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:01:36 crc kubenswrapper[4741]: E0929 22:01:36.087506 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:01:51 crc kubenswrapper[4741]: I0929 22:01:51.087560 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:01:51 crc kubenswrapper[4741]: E0929 22:01:51.089154 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:02:03 crc kubenswrapper[4741]: I0929 22:02:03.090354 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:02:03 crc kubenswrapper[4741]: E0929 22:02:03.091170 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:02:18 crc kubenswrapper[4741]: I0929 22:02:18.087224 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:02:18 crc kubenswrapper[4741]: E0929 22:02:18.091341 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:02:33 crc kubenswrapper[4741]: I0929 22:02:33.086923 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:02:33 crc kubenswrapper[4741]: E0929 22:02:33.088834 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:02:34 crc kubenswrapper[4741]: I0929 22:02:34.096214 4741 scope.go:117] "RemoveContainer" containerID="d3e60fb8e9ffa9895a8f9f9932b18b3fd6ad6c0b655dcb0ddcac2c63545f8b8d" Sep 29 22:02:45 crc kubenswrapper[4741]: I0929 22:02:45.086257 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:02:45 crc kubenswrapper[4741]: E0929 22:02:45.088335 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:02:56 crc kubenswrapper[4741]: I0929 22:02:56.086062 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:02:56 crc kubenswrapper[4741]: E0929 22:02:56.087012 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:03:10 crc kubenswrapper[4741]: I0929 22:03:10.087839 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:03:10 crc kubenswrapper[4741]: E0929 22:03:10.088867 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:03:22 crc kubenswrapper[4741]: I0929 22:03:22.087330 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:03:22 crc kubenswrapper[4741]: E0929 22:03:22.089815 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:03:35 crc kubenswrapper[4741]: I0929 22:03:35.087105 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:03:35 crc kubenswrapper[4741]: E0929 22:03:35.087892 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:03:45 crc kubenswrapper[4741]: I0929 22:03:45.515716 4741 generic.go:334] "Generic (PLEG): container finished" podID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerID="7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260" exitCode=0 Sep 29 22:03:45 crc kubenswrapper[4741]: I0929 22:03:45.515789 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4m44r/must-gather-dqv2x" event={"ID":"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7","Type":"ContainerDied","Data":"7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260"} Sep 29 22:03:45 crc kubenswrapper[4741]: I0929 22:03:45.517414 4741 scope.go:117] "RemoveContainer" containerID="7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260" Sep 29 22:03:46 crc kubenswrapper[4741]: I0929 22:03:46.429504 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4m44r_must-gather-dqv2x_8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7/gather/0.log" Sep 29 22:03:50 crc kubenswrapper[4741]: I0929 22:03:50.087722 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:03:50 crc kubenswrapper[4741]: E0929 22:03:50.089531 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:03:55 crc kubenswrapper[4741]: I0929 22:03:55.830911 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4m44r/must-gather-dqv2x"] Sep 29 22:03:55 crc kubenswrapper[4741]: I0929 22:03:55.831633 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4m44r/must-gather-dqv2x" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerName="copy" containerID="cri-o://116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32" gracePeriod=2 Sep 29 22:03:55 crc kubenswrapper[4741]: I0929 22:03:55.842940 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4m44r/must-gather-dqv2x"] Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.335714 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4m44r_must-gather-dqv2x_8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7/copy/0.log" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.336560 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.497170 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xk9c\" (UniqueName: \"kubernetes.io/projected/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-kube-api-access-4xk9c\") pod \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.497337 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-must-gather-output\") pod \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\" (UID: \"8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7\") " Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.522270 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-kube-api-access-4xk9c" (OuterVolumeSpecName: "kube-api-access-4xk9c") pod "8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" (UID: "8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7"). InnerVolumeSpecName "kube-api-access-4xk9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.600007 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xk9c\" (UniqueName: \"kubernetes.io/projected/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-kube-api-access-4xk9c\") on node \"crc\" DevicePath \"\"" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.638865 4741 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4m44r_must-gather-dqv2x_8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7/copy/0.log" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.639511 4741 generic.go:334] "Generic (PLEG): container finished" podID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerID="116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32" exitCode=143 Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.639600 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4m44r/must-gather-dqv2x" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.639636 4741 scope.go:117] "RemoveContainer" containerID="116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.665499 4741 scope.go:117] "RemoveContainer" containerID="7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.751729 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" (UID: "8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.754455 4741 scope.go:117] "RemoveContainer" containerID="116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32" Sep 29 22:03:56 crc kubenswrapper[4741]: E0929 22:03:56.754996 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32\": container with ID starting with 116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32 not found: ID does not exist" containerID="116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.755037 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32"} err="failed to get container status \"116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32\": rpc error: code = NotFound desc = could not find container \"116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32\": container with ID starting with 116ff7e808c30db85e660b3b155a95b97cdbd1420f98f460ccecbfa02d958c32 not found: ID does not exist" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.755063 4741 scope.go:117] "RemoveContainer" containerID="7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260" Sep 29 22:03:56 crc kubenswrapper[4741]: E0929 22:03:56.755568 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260\": container with ID starting with 7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260 not found: ID does not exist" containerID="7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.755616 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260"} err="failed to get container status \"7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260\": rpc error: code = NotFound desc = could not find container \"7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260\": container with ID starting with 7c62b8483148eddc4bd6add569791c27312cc38c11723ca2039c3319390ac260 not found: ID does not exist" Sep 29 22:03:56 crc kubenswrapper[4741]: I0929 22:03:56.805660 4741 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 29 22:03:57 crc kubenswrapper[4741]: I0929 22:03:57.099720 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" path="/var/lib/kubelet/pods/8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7/volumes" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.509142 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wqv4d"] Sep 29 22:04:00 crc kubenswrapper[4741]: E0929 22:04:00.510211 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b01432f5-24f4-45f9-81f9-32f05dea806b" containerName="keystone-cron" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.510228 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="b01432f5-24f4-45f9-81f9-32f05dea806b" containerName="keystone-cron" Sep 29 22:04:00 crc kubenswrapper[4741]: E0929 22:04:00.510244 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerName="gather" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.510252 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerName="gather" Sep 29 22:04:00 crc kubenswrapper[4741]: E0929 22:04:00.510300 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerName="copy" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.510308 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerName="copy" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.510576 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerName="gather" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.510614 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbe9a0e-3fb2-4457-acf2-aa1c08be76b7" containerName="copy" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.510631 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="b01432f5-24f4-45f9-81f9-32f05dea806b" containerName="keystone-cron" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.512520 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.520977 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqv4d"] Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.592782 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-utilities\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.592969 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-catalog-content\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.593282 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zscsm\" (UniqueName: \"kubernetes.io/projected/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-kube-api-access-zscsm\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.695280 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-catalog-content\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.695490 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zscsm\" (UniqueName: \"kubernetes.io/projected/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-kube-api-access-zscsm\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.695585 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-utilities\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.695942 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-catalog-content\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.696082 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-utilities\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.716474 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zscsm\" (UniqueName: \"kubernetes.io/projected/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-kube-api-access-zscsm\") pod \"redhat-marketplace-wqv4d\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:00 crc kubenswrapper[4741]: I0929 22:04:00.834823 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:01 crc kubenswrapper[4741]: I0929 22:04:01.086509 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:04:01 crc kubenswrapper[4741]: E0929 22:04:01.087081 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:04:01 crc kubenswrapper[4741]: I0929 22:04:01.321299 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqv4d"] Sep 29 22:04:01 crc kubenswrapper[4741]: I0929 22:04:01.710539 4741 generic.go:334] "Generic (PLEG): container finished" podID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerID="371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66" exitCode=0 Sep 29 22:04:01 crc kubenswrapper[4741]: I0929 22:04:01.710588 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqv4d" event={"ID":"ac80363c-411d-4d48-bb8d-e6ec0f0d2473","Type":"ContainerDied","Data":"371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66"} Sep 29 22:04:01 crc kubenswrapper[4741]: I0929 22:04:01.710969 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqv4d" event={"ID":"ac80363c-411d-4d48-bb8d-e6ec0f0d2473","Type":"ContainerStarted","Data":"17eb55b16226ef07b1459cd82f39a8164fbeba3d883d34fe0b1c5e1bae459e6f"} Sep 29 22:04:01 crc kubenswrapper[4741]: I0929 22:04:01.712458 4741 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 22:04:02 crc kubenswrapper[4741]: I0929 22:04:02.721328 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqv4d" event={"ID":"ac80363c-411d-4d48-bb8d-e6ec0f0d2473","Type":"ContainerStarted","Data":"778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe"} Sep 29 22:04:03 crc kubenswrapper[4741]: I0929 22:04:03.731729 4741 generic.go:334] "Generic (PLEG): container finished" podID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerID="778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe" exitCode=0 Sep 29 22:04:03 crc kubenswrapper[4741]: I0929 22:04:03.732569 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqv4d" event={"ID":"ac80363c-411d-4d48-bb8d-e6ec0f0d2473","Type":"ContainerDied","Data":"778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe"} Sep 29 22:04:04 crc kubenswrapper[4741]: I0929 22:04:04.742115 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqv4d" event={"ID":"ac80363c-411d-4d48-bb8d-e6ec0f0d2473","Type":"ContainerStarted","Data":"47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a"} Sep 29 22:04:04 crc kubenswrapper[4741]: I0929 22:04:04.765175 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wqv4d" podStartSLOduration=2.200128899 podStartE2EDuration="4.765157311s" podCreationTimestamp="2025-09-29 22:04:00 +0000 UTC" firstStartedPulling="2025-09-29 22:04:01.712224095 +0000 UTC m=+10483.360013427" lastFinishedPulling="2025-09-29 22:04:04.277252507 +0000 UTC m=+10485.925041839" observedRunningTime="2025-09-29 22:04:04.757625634 +0000 UTC m=+10486.405414966" watchObservedRunningTime="2025-09-29 22:04:04.765157311 +0000 UTC m=+10486.412946643" Sep 29 22:04:10 crc kubenswrapper[4741]: I0929 22:04:10.836597 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:10 crc kubenswrapper[4741]: I0929 22:04:10.837921 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:10 crc kubenswrapper[4741]: I0929 22:04:10.894911 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:11 crc kubenswrapper[4741]: I0929 22:04:11.871987 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:11 crc kubenswrapper[4741]: I0929 22:04:11.933661 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqv4d"] Sep 29 22:04:13 crc kubenswrapper[4741]: I0929 22:04:13.839103 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wqv4d" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="registry-server" containerID="cri-o://47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a" gracePeriod=2 Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.412001 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.522251 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-utilities\") pod \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.522384 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zscsm\" (UniqueName: \"kubernetes.io/projected/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-kube-api-access-zscsm\") pod \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.522510 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-catalog-content\") pod \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\" (UID: \"ac80363c-411d-4d48-bb8d-e6ec0f0d2473\") " Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.524593 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-utilities" (OuterVolumeSpecName: "utilities") pod "ac80363c-411d-4d48-bb8d-e6ec0f0d2473" (UID: "ac80363c-411d-4d48-bb8d-e6ec0f0d2473"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.530451 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-kube-api-access-zscsm" (OuterVolumeSpecName: "kube-api-access-zscsm") pod "ac80363c-411d-4d48-bb8d-e6ec0f0d2473" (UID: "ac80363c-411d-4d48-bb8d-e6ec0f0d2473"). InnerVolumeSpecName "kube-api-access-zscsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.543043 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac80363c-411d-4d48-bb8d-e6ec0f0d2473" (UID: "ac80363c-411d-4d48-bb8d-e6ec0f0d2473"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.624455 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zscsm\" (UniqueName: \"kubernetes.io/projected/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-kube-api-access-zscsm\") on node \"crc\" DevicePath \"\"" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.624679 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.624771 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac80363c-411d-4d48-bb8d-e6ec0f0d2473-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.852095 4741 generic.go:334] "Generic (PLEG): container finished" podID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerID="47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a" exitCode=0 Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.852148 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqv4d" event={"ID":"ac80363c-411d-4d48-bb8d-e6ec0f0d2473","Type":"ContainerDied","Data":"47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a"} Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.852178 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqv4d" event={"ID":"ac80363c-411d-4d48-bb8d-e6ec0f0d2473","Type":"ContainerDied","Data":"17eb55b16226ef07b1459cd82f39a8164fbeba3d883d34fe0b1c5e1bae459e6f"} Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.852196 4741 scope.go:117] "RemoveContainer" containerID="47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.852331 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqv4d" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.881454 4741 scope.go:117] "RemoveContainer" containerID="778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.896880 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqv4d"] Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.915753 4741 scope.go:117] "RemoveContainer" containerID="371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.927446 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqv4d"] Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.957277 4741 scope.go:117] "RemoveContainer" containerID="47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a" Sep 29 22:04:14 crc kubenswrapper[4741]: E0929 22:04:14.957837 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a\": container with ID starting with 47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a not found: ID does not exist" containerID="47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.957896 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a"} err="failed to get container status \"47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a\": rpc error: code = NotFound desc = could not find container \"47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a\": container with ID starting with 47564788c546f7f6f547bfa6ce9cdb98909d78777df12eccaa8065d028f5ec6a not found: ID does not exist" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.957983 4741 scope.go:117] "RemoveContainer" containerID="778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe" Sep 29 22:04:14 crc kubenswrapper[4741]: E0929 22:04:14.958542 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe\": container with ID starting with 778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe not found: ID does not exist" containerID="778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.958582 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe"} err="failed to get container status \"778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe\": rpc error: code = NotFound desc = could not find container \"778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe\": container with ID starting with 778865ec4d4ec559638bbbdf78b18b3d59b189646d60a3410954fe27acc120fe not found: ID does not exist" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.958614 4741 scope.go:117] "RemoveContainer" containerID="371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66" Sep 29 22:04:14 crc kubenswrapper[4741]: E0929 22:04:14.959043 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66\": container with ID starting with 371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66 not found: ID does not exist" containerID="371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66" Sep 29 22:04:14 crc kubenswrapper[4741]: I0929 22:04:14.959103 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66"} err="failed to get container status \"371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66\": rpc error: code = NotFound desc = could not find container \"371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66\": container with ID starting with 371c80e7563f7d0f62ffa76b47639a1cb2988a58a436f9357fa6358a782a9e66 not found: ID does not exist" Sep 29 22:04:15 crc kubenswrapper[4741]: I0929 22:04:15.100767 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" path="/var/lib/kubelet/pods/ac80363c-411d-4d48-bb8d-e6ec0f0d2473/volumes" Sep 29 22:04:16 crc kubenswrapper[4741]: I0929 22:04:16.087085 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:04:16 crc kubenswrapper[4741]: E0929 22:04:16.088302 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:04:31 crc kubenswrapper[4741]: I0929 22:04:31.086327 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:04:31 crc kubenswrapper[4741]: E0929 22:04:31.087262 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:04:34 crc kubenswrapper[4741]: I0929 22:04:34.194701 4741 scope.go:117] "RemoveContainer" containerID="81a0debb7e1393db51cbfb2547eebe48f3b9601d667609e7ec7fad56ae3798d0" Sep 29 22:04:42 crc kubenswrapper[4741]: I0929 22:04:42.086302 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:04:42 crc kubenswrapper[4741]: E0929 22:04:42.087273 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:04:45 crc kubenswrapper[4741]: I0929 22:04:45.877052 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9lnpl"] Sep 29 22:04:45 crc kubenswrapper[4741]: E0929 22:04:45.878770 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="extract-utilities" Sep 29 22:04:45 crc kubenswrapper[4741]: I0929 22:04:45.878793 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="extract-utilities" Sep 29 22:04:45 crc kubenswrapper[4741]: E0929 22:04:45.878829 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="extract-content" Sep 29 22:04:45 crc kubenswrapper[4741]: I0929 22:04:45.878840 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="extract-content" Sep 29 22:04:45 crc kubenswrapper[4741]: E0929 22:04:45.878858 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="registry-server" Sep 29 22:04:45 crc kubenswrapper[4741]: I0929 22:04:45.878866 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="registry-server" Sep 29 22:04:45 crc kubenswrapper[4741]: I0929 22:04:45.879193 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac80363c-411d-4d48-bb8d-e6ec0f0d2473" containerName="registry-server" Sep 29 22:04:45 crc kubenswrapper[4741]: I0929 22:04:45.881460 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:45 crc kubenswrapper[4741]: I0929 22:04:45.910980 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lnpl"] Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.050883 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-catalog-content\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.051510 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-utilities\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.051574 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlx5t\" (UniqueName: \"kubernetes.io/projected/4c23877f-ff70-455e-ac68-0813badadc17-kube-api-access-hlx5t\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.155407 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-catalog-content\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.155628 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-utilities\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.155724 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlx5t\" (UniqueName: \"kubernetes.io/projected/4c23877f-ff70-455e-ac68-0813badadc17-kube-api-access-hlx5t\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.156177 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-catalog-content\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.156506 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-utilities\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.183232 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlx5t\" (UniqueName: \"kubernetes.io/projected/4c23877f-ff70-455e-ac68-0813badadc17-kube-api-access-hlx5t\") pod \"certified-operators-9lnpl\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.219121 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:46 crc kubenswrapper[4741]: I0929 22:04:46.828087 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lnpl"] Sep 29 22:04:47 crc kubenswrapper[4741]: I0929 22:04:47.219070 4741 generic.go:334] "Generic (PLEG): container finished" podID="4c23877f-ff70-455e-ac68-0813badadc17" containerID="060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21" exitCode=0 Sep 29 22:04:47 crc kubenswrapper[4741]: I0929 22:04:47.219132 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lnpl" event={"ID":"4c23877f-ff70-455e-ac68-0813badadc17","Type":"ContainerDied","Data":"060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21"} Sep 29 22:04:47 crc kubenswrapper[4741]: I0929 22:04:47.219513 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lnpl" event={"ID":"4c23877f-ff70-455e-ac68-0813badadc17","Type":"ContainerStarted","Data":"26fced899e5f9a386f3a5c42042c052de0295671bdccac2c9c5f947469020867"} Sep 29 22:04:48 crc kubenswrapper[4741]: I0929 22:04:48.232462 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lnpl" event={"ID":"4c23877f-ff70-455e-ac68-0813badadc17","Type":"ContainerStarted","Data":"3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84"} Sep 29 22:04:49 crc kubenswrapper[4741]: I0929 22:04:49.243852 4741 generic.go:334] "Generic (PLEG): container finished" podID="4c23877f-ff70-455e-ac68-0813badadc17" containerID="3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84" exitCode=0 Sep 29 22:04:49 crc kubenswrapper[4741]: I0929 22:04:49.243954 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lnpl" event={"ID":"4c23877f-ff70-455e-ac68-0813badadc17","Type":"ContainerDied","Data":"3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84"} Sep 29 22:04:50 crc kubenswrapper[4741]: I0929 22:04:50.260238 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lnpl" event={"ID":"4c23877f-ff70-455e-ac68-0813badadc17","Type":"ContainerStarted","Data":"45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0"} Sep 29 22:04:50 crc kubenswrapper[4741]: I0929 22:04:50.287609 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9lnpl" podStartSLOduration=2.780754503 podStartE2EDuration="5.287578993s" podCreationTimestamp="2025-09-29 22:04:45 +0000 UTC" firstStartedPulling="2025-09-29 22:04:47.223092623 +0000 UTC m=+10528.870881955" lastFinishedPulling="2025-09-29 22:04:49.729917113 +0000 UTC m=+10531.377706445" observedRunningTime="2025-09-29 22:04:50.283498964 +0000 UTC m=+10531.931288326" watchObservedRunningTime="2025-09-29 22:04:50.287578993 +0000 UTC m=+10531.935368325" Sep 29 22:04:55 crc kubenswrapper[4741]: I0929 22:04:55.095358 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:04:55 crc kubenswrapper[4741]: E0929 22:04:55.096277 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:04:56 crc kubenswrapper[4741]: I0929 22:04:56.220271 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:56 crc kubenswrapper[4741]: I0929 22:04:56.220341 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:56 crc kubenswrapper[4741]: I0929 22:04:56.276795 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:56 crc kubenswrapper[4741]: I0929 22:04:56.408732 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:56 crc kubenswrapper[4741]: I0929 22:04:56.516504 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lnpl"] Sep 29 22:04:58 crc kubenswrapper[4741]: I0929 22:04:58.380465 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9lnpl" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="registry-server" containerID="cri-o://45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0" gracePeriod=2 Sep 29 22:04:58 crc kubenswrapper[4741]: I0929 22:04:58.898137 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:58 crc kubenswrapper[4741]: I0929 22:04:58.943309 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlx5t\" (UniqueName: \"kubernetes.io/projected/4c23877f-ff70-455e-ac68-0813badadc17-kube-api-access-hlx5t\") pod \"4c23877f-ff70-455e-ac68-0813badadc17\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " Sep 29 22:04:58 crc kubenswrapper[4741]: I0929 22:04:58.943479 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-catalog-content\") pod \"4c23877f-ff70-455e-ac68-0813badadc17\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " Sep 29 22:04:58 crc kubenswrapper[4741]: I0929 22:04:58.943728 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-utilities\") pod \"4c23877f-ff70-455e-ac68-0813badadc17\" (UID: \"4c23877f-ff70-455e-ac68-0813badadc17\") " Sep 29 22:04:58 crc kubenswrapper[4741]: I0929 22:04:58.944742 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-utilities" (OuterVolumeSpecName: "utilities") pod "4c23877f-ff70-455e-ac68-0813badadc17" (UID: "4c23877f-ff70-455e-ac68-0813badadc17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 22:04:58 crc kubenswrapper[4741]: I0929 22:04:58.949505 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c23877f-ff70-455e-ac68-0813badadc17-kube-api-access-hlx5t" (OuterVolumeSpecName: "kube-api-access-hlx5t") pod "4c23877f-ff70-455e-ac68-0813badadc17" (UID: "4c23877f-ff70-455e-ac68-0813badadc17"). InnerVolumeSpecName "kube-api-access-hlx5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.047775 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlx5t\" (UniqueName: \"kubernetes.io/projected/4c23877f-ff70-455e-ac68-0813badadc17-kube-api-access-hlx5t\") on node \"crc\" DevicePath \"\"" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.047835 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.089027 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c23877f-ff70-455e-ac68-0813badadc17" (UID: "4c23877f-ff70-455e-ac68-0813badadc17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.150382 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c23877f-ff70-455e-ac68-0813badadc17-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.400430 4741 generic.go:334] "Generic (PLEG): container finished" podID="4c23877f-ff70-455e-ac68-0813badadc17" containerID="45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0" exitCode=0 Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.400528 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lnpl" event={"ID":"4c23877f-ff70-455e-ac68-0813badadc17","Type":"ContainerDied","Data":"45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0"} Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.400578 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lnpl" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.400794 4741 scope.go:117] "RemoveContainer" containerID="45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.400774 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lnpl" event={"ID":"4c23877f-ff70-455e-ac68-0813badadc17","Type":"ContainerDied","Data":"26fced899e5f9a386f3a5c42042c052de0295671bdccac2c9c5f947469020867"} Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.423600 4741 scope.go:117] "RemoveContainer" containerID="3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.451207 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lnpl"] Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.461873 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9lnpl"] Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.466161 4741 scope.go:117] "RemoveContainer" containerID="060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.520659 4741 scope.go:117] "RemoveContainer" containerID="45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0" Sep 29 22:04:59 crc kubenswrapper[4741]: E0929 22:04:59.521231 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0\": container with ID starting with 45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0 not found: ID does not exist" containerID="45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.521290 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0"} err="failed to get container status \"45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0\": rpc error: code = NotFound desc = could not find container \"45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0\": container with ID starting with 45cab2280391c8174cd1d6321ece990786bd3318584dd0b0f4da5fad589f75e0 not found: ID does not exist" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.521321 4741 scope.go:117] "RemoveContainer" containerID="3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84" Sep 29 22:04:59 crc kubenswrapper[4741]: E0929 22:04:59.521907 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84\": container with ID starting with 3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84 not found: ID does not exist" containerID="3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.521951 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84"} err="failed to get container status \"3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84\": rpc error: code = NotFound desc = could not find container \"3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84\": container with ID starting with 3bf42e59257e1145adc19dae0aa9f4d921a3052a1edd5c710ea24d15fda9ab84 not found: ID does not exist" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.521979 4741 scope.go:117] "RemoveContainer" containerID="060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21" Sep 29 22:04:59 crc kubenswrapper[4741]: E0929 22:04:59.522262 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21\": container with ID starting with 060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21 not found: ID does not exist" containerID="060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21" Sep 29 22:04:59 crc kubenswrapper[4741]: I0929 22:04:59.522297 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21"} err="failed to get container status \"060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21\": rpc error: code = NotFound desc = could not find container \"060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21\": container with ID starting with 060da32fbaaeed0c6d2c0f42b3e01a96e96ba98b935a985d6c86e202c94bcd21 not found: ID does not exist" Sep 29 22:05:01 crc kubenswrapper[4741]: I0929 22:05:01.110110 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c23877f-ff70-455e-ac68-0813badadc17" path="/var/lib/kubelet/pods/4c23877f-ff70-455e-ac68-0813badadc17/volumes" Sep 29 22:05:06 crc kubenswrapper[4741]: I0929 22:05:06.086270 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:05:06 crc kubenswrapper[4741]: E0929 22:05:06.087039 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.655884 4741 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q5788"] Sep 29 22:05:18 crc kubenswrapper[4741]: E0929 22:05:18.660062 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="extract-content" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.660101 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="extract-content" Sep 29 22:05:18 crc kubenswrapper[4741]: E0929 22:05:18.660146 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="registry-server" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.660154 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="registry-server" Sep 29 22:05:18 crc kubenswrapper[4741]: E0929 22:05:18.660181 4741 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="extract-utilities" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.660189 4741 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="extract-utilities" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.660561 4741 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c23877f-ff70-455e-ac68-0813badadc17" containerName="registry-server" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.663765 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.676348 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5788"] Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.839902 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-utilities\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.840155 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-catalog-content\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.840181 4741 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpbc\" (UniqueName: \"kubernetes.io/projected/0c9469a9-cfdf-4242-a7b6-bb0868f97264-kube-api-access-kwpbc\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.942801 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-catalog-content\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.942873 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpbc\" (UniqueName: \"kubernetes.io/projected/0c9469a9-cfdf-4242-a7b6-bb0868f97264-kube-api-access-kwpbc\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.943443 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-catalog-content\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.943457 4741 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-utilities\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.943933 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-utilities\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.963478 4741 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpbc\" (UniqueName: \"kubernetes.io/projected/0c9469a9-cfdf-4242-a7b6-bb0868f97264-kube-api-access-kwpbc\") pod \"redhat-operators-q5788\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:18 crc kubenswrapper[4741]: I0929 22:05:18.999542 4741 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:19 crc kubenswrapper[4741]: I0929 22:05:19.517455 4741 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5788"] Sep 29 22:05:19 crc kubenswrapper[4741]: I0929 22:05:19.656696 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5788" event={"ID":"0c9469a9-cfdf-4242-a7b6-bb0868f97264","Type":"ContainerStarted","Data":"5a8ef255241ae814c9e2f2bfbfe0f20f485b2a74ea8aa379366a231fd34db700"} Sep 29 22:05:20 crc kubenswrapper[4741]: I0929 22:05:20.667547 4741 generic.go:334] "Generic (PLEG): container finished" podID="0c9469a9-cfdf-4242-a7b6-bb0868f97264" containerID="01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859" exitCode=0 Sep 29 22:05:20 crc kubenswrapper[4741]: I0929 22:05:20.667735 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5788" event={"ID":"0c9469a9-cfdf-4242-a7b6-bb0868f97264","Type":"ContainerDied","Data":"01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859"} Sep 29 22:05:21 crc kubenswrapper[4741]: I0929 22:05:21.086652 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:05:21 crc kubenswrapper[4741]: E0929 22:05:21.086937 4741 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rtxqk_openshift-machine-config-operator(b2f20aa2-390a-494e-aed0-0dfd0e031d3d)\"" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" podUID="b2f20aa2-390a-494e-aed0-0dfd0e031d3d" Sep 29 22:05:22 crc kubenswrapper[4741]: I0929 22:05:22.691202 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5788" event={"ID":"0c9469a9-cfdf-4242-a7b6-bb0868f97264","Type":"ContainerStarted","Data":"e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63"} Sep 29 22:05:24 crc kubenswrapper[4741]: E0929 22:05:24.567979 4741 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c9469a9_cfdf_4242_a7b6_bb0868f97264.slice/crio-conmon-e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c9469a9_cfdf_4242_a7b6_bb0868f97264.slice/crio-e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63.scope\": RecentStats: unable to find data in memory cache]" Sep 29 22:05:24 crc kubenswrapper[4741]: I0929 22:05:24.711625 4741 generic.go:334] "Generic (PLEG): container finished" podID="0c9469a9-cfdf-4242-a7b6-bb0868f97264" containerID="e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63" exitCode=0 Sep 29 22:05:24 crc kubenswrapper[4741]: I0929 22:05:24.711672 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5788" event={"ID":"0c9469a9-cfdf-4242-a7b6-bb0868f97264","Type":"ContainerDied","Data":"e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63"} Sep 29 22:05:25 crc kubenswrapper[4741]: I0929 22:05:25.721739 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5788" event={"ID":"0c9469a9-cfdf-4242-a7b6-bb0868f97264","Type":"ContainerStarted","Data":"176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf"} Sep 29 22:05:25 crc kubenswrapper[4741]: I0929 22:05:25.745547 4741 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q5788" podStartSLOduration=3.317830462 podStartE2EDuration="7.745524182s" podCreationTimestamp="2025-09-29 22:05:18 +0000 UTC" firstStartedPulling="2025-09-29 22:05:20.669790658 +0000 UTC m=+10562.317579980" lastFinishedPulling="2025-09-29 22:05:25.097484368 +0000 UTC m=+10566.745273700" observedRunningTime="2025-09-29 22:05:25.737295403 +0000 UTC m=+10567.385084735" watchObservedRunningTime="2025-09-29 22:05:25.745524182 +0000 UTC m=+10567.393313514" Sep 29 22:05:29 crc kubenswrapper[4741]: I0929 22:05:28.999692 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:29 crc kubenswrapper[4741]: I0929 22:05:29.000278 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:30 crc kubenswrapper[4741]: I0929 22:05:30.048343 4741 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q5788" podUID="0c9469a9-cfdf-4242-a7b6-bb0868f97264" containerName="registry-server" probeResult="failure" output=< Sep 29 22:05:30 crc kubenswrapper[4741]: timeout: failed to connect service ":50051" within 1s Sep 29 22:05:30 crc kubenswrapper[4741]: > Sep 29 22:05:33 crc kubenswrapper[4741]: I0929 22:05:33.085740 4741 scope.go:117] "RemoveContainer" containerID="1d2338b07b1d3a9933a1a16aac082fb8ff2a6f68852236ebeb53d4542e42a45a" Sep 29 22:05:33 crc kubenswrapper[4741]: I0929 22:05:33.797359 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rtxqk" event={"ID":"b2f20aa2-390a-494e-aed0-0dfd0e031d3d","Type":"ContainerStarted","Data":"dcfe5d44f16e093eada3372332a6179269a6ff9e1cd7296b231f3c887cd4217b"} Sep 29 22:05:39 crc kubenswrapper[4741]: I0929 22:05:39.056728 4741 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:39 crc kubenswrapper[4741]: I0929 22:05:39.104310 4741 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:39 crc kubenswrapper[4741]: I0929 22:05:39.290546 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5788"] Sep 29 22:05:40 crc kubenswrapper[4741]: I0929 22:05:40.878233 4741 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q5788" podUID="0c9469a9-cfdf-4242-a7b6-bb0868f97264" containerName="registry-server" containerID="cri-o://176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf" gracePeriod=2 Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.429022 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.541348 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-catalog-content\") pod \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.541418 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwpbc\" (UniqueName: \"kubernetes.io/projected/0c9469a9-cfdf-4242-a7b6-bb0868f97264-kube-api-access-kwpbc\") pod \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.541452 4741 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-utilities\") pod \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\" (UID: \"0c9469a9-cfdf-4242-a7b6-bb0868f97264\") " Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.545328 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-utilities" (OuterVolumeSpecName: "utilities") pod "0c9469a9-cfdf-4242-a7b6-bb0868f97264" (UID: "0c9469a9-cfdf-4242-a7b6-bb0868f97264"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.562578 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9469a9-cfdf-4242-a7b6-bb0868f97264-kube-api-access-kwpbc" (OuterVolumeSpecName: "kube-api-access-kwpbc") pod "0c9469a9-cfdf-4242-a7b6-bb0868f97264" (UID: "0c9469a9-cfdf-4242-a7b6-bb0868f97264"). InnerVolumeSpecName "kube-api-access-kwpbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.626084 4741 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c9469a9-cfdf-4242-a7b6-bb0868f97264" (UID: "0c9469a9-cfdf-4242-a7b6-bb0868f97264"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.643422 4741 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.643461 4741 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwpbc\" (UniqueName: \"kubernetes.io/projected/0c9469a9-cfdf-4242-a7b6-bb0868f97264-kube-api-access-kwpbc\") on node \"crc\" DevicePath \"\"" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.643475 4741 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9469a9-cfdf-4242-a7b6-bb0868f97264-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.893424 4741 generic.go:334] "Generic (PLEG): container finished" podID="0c9469a9-cfdf-4242-a7b6-bb0868f97264" containerID="176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf" exitCode=0 Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.893469 4741 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5788" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.893495 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5788" event={"ID":"0c9469a9-cfdf-4242-a7b6-bb0868f97264","Type":"ContainerDied","Data":"176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf"} Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.893606 4741 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5788" event={"ID":"0c9469a9-cfdf-4242-a7b6-bb0868f97264","Type":"ContainerDied","Data":"5a8ef255241ae814c9e2f2bfbfe0f20f485b2a74ea8aa379366a231fd34db700"} Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.893650 4741 scope.go:117] "RemoveContainer" containerID="176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.941332 4741 scope.go:117] "RemoveContainer" containerID="e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63" Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.945316 4741 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5788"] Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.956054 4741 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q5788"] Sep 29 22:05:41 crc kubenswrapper[4741]: I0929 22:05:41.996420 4741 scope.go:117] "RemoveContainer" containerID="01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859" Sep 29 22:05:42 crc kubenswrapper[4741]: I0929 22:05:42.044477 4741 scope.go:117] "RemoveContainer" containerID="176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf" Sep 29 22:05:42 crc kubenswrapper[4741]: E0929 22:05:42.046139 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf\": container with ID starting with 176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf not found: ID does not exist" containerID="176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf" Sep 29 22:05:42 crc kubenswrapper[4741]: I0929 22:05:42.046193 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf"} err="failed to get container status \"176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf\": rpc error: code = NotFound desc = could not find container \"176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf\": container with ID starting with 176d4ceb7cab7a4712caac748549bb64a2cb27d8f823cb8e3cd226f8e62d00bf not found: ID does not exist" Sep 29 22:05:42 crc kubenswrapper[4741]: I0929 22:05:42.046230 4741 scope.go:117] "RemoveContainer" containerID="e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63" Sep 29 22:05:42 crc kubenswrapper[4741]: E0929 22:05:42.046777 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63\": container with ID starting with e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63 not found: ID does not exist" containerID="e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63" Sep 29 22:05:42 crc kubenswrapper[4741]: I0929 22:05:42.046812 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63"} err="failed to get container status \"e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63\": rpc error: code = NotFound desc = could not find container \"e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63\": container with ID starting with e966495753f7c2ae138dc4c9fa6244a191d775754f5aa27816f83c13f4f62e63 not found: ID does not exist" Sep 29 22:05:42 crc kubenswrapper[4741]: I0929 22:05:42.046834 4741 scope.go:117] "RemoveContainer" containerID="01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859" Sep 29 22:05:42 crc kubenswrapper[4741]: E0929 22:05:42.047144 4741 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859\": container with ID starting with 01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859 not found: ID does not exist" containerID="01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859" Sep 29 22:05:42 crc kubenswrapper[4741]: I0929 22:05:42.047175 4741 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859"} err="failed to get container status \"01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859\": rpc error: code = NotFound desc = could not find container \"01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859\": container with ID starting with 01547e202ac1815c82e16244945857075da99051654e1a3a5665ebf130811859 not found: ID does not exist" Sep 29 22:05:43 crc kubenswrapper[4741]: I0929 22:05:43.100324 4741 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9469a9-cfdf-4242-a7b6-bb0868f97264" path="/var/lib/kubelet/pods/0c9469a9-cfdf-4242-a7b6-bb0868f97264/volumes"